| """Test evaluating expressions repeatedly comparing lldb against gdb.""" |
| |
| from __future__ import print_function |
| |
| |
| import os |
| import sys |
| import lldb |
| from lldbsuite.test.lldbbench import BenchBase |
| from lldbsuite.test.decorators import * |
| from lldbsuite.test.lldbtest import * |
| from lldbsuite.test import configuration |
| from lldbsuite.test import lldbutil |
| |
| |
| class RepeatedExprsCase(BenchBase): |
| |
| mydir = TestBase.compute_mydir(__file__) |
| |
| def setUp(self): |
| BenchBase.setUp(self) |
| self.source = 'main.cpp' |
| self.line_to_break = line_number( |
| self.source, '// Set breakpoint here.') |
| self.lldb_avg = None |
| self.gdb_avg = None |
| self.count = 100 |
| |
| @benchmarks_test |
| @expectedFailureAll( |
| oslist=["windows"], |
| bugnumber="llvm.org/pr22274: need a pexpect replacement for windows") |
| def test_compare_lldb_to_gdb(self): |
| """Test repeated expressions with lldb vs. gdb.""" |
| self.build() |
| self.exe_name = 'a.out' |
| |
| print() |
| self.run_lldb_repeated_exprs(self.exe_name, self.count) |
| print("lldb benchmark:", self.stopwatch) |
| self.run_gdb_repeated_exprs(self.exe_name, self.count) |
| print("gdb benchmark:", self.stopwatch) |
| print("lldb_avg/gdb_avg: %f" % (self.lldb_avg / self.gdb_avg)) |
| |
| def run_lldb_repeated_exprs(self, exe_name, count): |
| import pexpect |
| exe = self.getBuildArtifact(exe_name) |
| |
| # Set self.child_prompt, which is "(lldb) ". |
| self.child_prompt = '(lldb) ' |
| prompt = self.child_prompt |
| |
| # So that the child gets torn down after the test. |
| self.child = pexpect.spawn( |
| '%s %s %s' % |
| (lldbtest_config.lldbExec, self.lldbOption, exe)) |
| child = self.child |
| |
| # Turn on logging for what the child sends back. |
| if self.TraceOn(): |
| child.logfile_read = sys.stdout |
| |
| child.expect_exact(prompt) |
| child.sendline( |
| 'breakpoint set -f %s -l %d' % |
| (self.source, self.line_to_break)) |
| child.expect_exact(prompt) |
| child.sendline('run') |
| child.expect_exact(prompt) |
| expr_cmd1 = 'expr ptr[j]->point.x' |
| expr_cmd2 = 'expr ptr[j]->point.y' |
| |
| # Reset the stopwatch now. |
| self.stopwatch.reset() |
| for i in range(count): |
| with self.stopwatch: |
| child.sendline(expr_cmd1) |
| child.expect_exact(prompt) |
| child.sendline(expr_cmd2) |
| child.expect_exact(prompt) |
| child.sendline('process continue') |
| child.expect_exact(prompt) |
| |
| child.sendline('quit') |
| try: |
| self.child.expect(pexpect.EOF) |
| except: |
| pass |
| |
| self.lldb_avg = self.stopwatch.avg() |
| if self.TraceOn(): |
| print("lldb expression benchmark:", str(self.stopwatch)) |
| self.child = None |
| |
| def run_gdb_repeated_exprs(self, exe_name, count): |
| import pexpect |
| exe = self.getBuildArtifact(exe_name) |
| |
| # Set self.child_prompt, which is "(gdb) ". |
| self.child_prompt = '(gdb) ' |
| prompt = self.child_prompt |
| |
| # So that the child gets torn down after the test. |
| self.child = pexpect.spawn('gdb --nx %s' % exe) |
| child = self.child |
| |
| # Turn on logging for what the child sends back. |
| if self.TraceOn(): |
| child.logfile_read = sys.stdout |
| |
| child.expect_exact(prompt) |
| child.sendline('break %s:%d' % (self.source, self.line_to_break)) |
| child.expect_exact(prompt) |
| child.sendline('run') |
| child.expect_exact(prompt) |
| expr_cmd1 = 'print ptr[j]->point.x' |
| expr_cmd2 = 'print ptr[j]->point.y' |
| |
| # Reset the stopwatch now. |
| self.stopwatch.reset() |
| for i in range(count): |
| with self.stopwatch: |
| child.sendline(expr_cmd1) |
| child.expect_exact(prompt) |
| child.sendline(expr_cmd2) |
| child.expect_exact(prompt) |
| child.sendline('continue') |
| child.expect_exact(prompt) |
| |
| child.sendline('quit') |
| child.expect_exact('The program is running. Exit anyway?') |
| child.sendline('y') |
| try: |
| self.child.expect(pexpect.EOF) |
| except: |
| pass |
| |
| self.gdb_avg = self.stopwatch.avg() |
| if self.TraceOn(): |
| print("gdb expression benchmark:", str(self.stopwatch)) |
| self.child = None |