| """ |
| Run the test suite using a separate process for each test file. |
| |
| Each test will run with a time limit of 10 minutes by default. |
| |
| Override the default time limit of 10 minutes by setting |
| the environment variable LLDB_TEST_TIMEOUT. |
| |
| E.g., export LLDB_TEST_TIMEOUT=10m |
| |
| Override the time limit for individual tests by setting |
| the environment variable LLDB_[TEST NAME]_TIMEOUT. |
| |
| E.g., export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=2m |
| |
| Set to "0" to run without time limit. |
| |
| E.g., export LLDB_TEST_TIMEOUT=0 |
| or export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=0 |
| |
| To collect core files for timed out tests, |
| do the following before running dosep.py |
| |
| OSX |
| ulimit -c unlimited |
| sudo sysctl -w kern.corefile=core.%P |
| |
| Linux: |
| ulimit -c unlimited |
| echo core.%p | sudo tee /proc/sys/kernel/core_pattern |
| """ |
| |
| from __future__ import absolute_import |
| from __future__ import print_function |
| |
| # system packages and modules |
| import asyncore |
| import distutils.version |
| import fnmatch |
| import multiprocessing |
| import multiprocessing.pool |
| import os |
| import platform |
| import re |
| import signal |
| import sys |
| import threading |
| |
| from six import StringIO |
| from six.moves import queue |
| |
| # Our packages and modules |
| import lldbsuite |
| import lldbsuite.support.seven as seven |
| |
| from . import configuration |
| from . import dotest_args |
| from lldbsuite.support import optional_with |
| from lldbsuite.test_event import dotest_channels |
| from lldbsuite.test_event.event_builder import EventBuilder |
| from lldbsuite.test_event import formatter |
| |
| from .test_runner import process_control |
| |
| # Status codes for running command with timeout. |
| eTimedOut, ePassed, eFailed = 124, 0, 1 |
| |
| g_session_dir = None |
| g_runner_context = None |
| output_lock = None |
| test_counter = None |
| total_tests = None |
| test_name_len = None |
| dotest_options = None |
| RESULTS_FORMATTER = None |
| RUNNER_PROCESS_ASYNC_MAP = None |
| RESULTS_LISTENER_CHANNEL = None |
| |
| """Contains an optional function pointer that can return the worker index |
| for the given thread/process calling it. Returns a 0-based index.""" |
| GET_WORKER_INDEX = None |
| |
| |
| def setup_global_variables( |
| lock, counter, total, name_len, options, worker_index_map): |
| global output_lock, test_counter, total_tests, test_name_len |
| global dotest_options |
| output_lock = lock |
| test_counter = counter |
| total_tests = total |
| test_name_len = name_len |
| dotest_options = options |
| |
| if worker_index_map is not None: |
| # We'll use the output lock for this to avoid sharing another lock. |
| # This won't be used much. |
| index_lock = lock |
| |
| def get_worker_index_use_pid(): |
| """Returns a 0-based, process-unique index for the worker.""" |
| pid = os.getpid() |
| with index_lock: |
| if pid not in worker_index_map: |
| worker_index_map[pid] = len(worker_index_map) |
| return worker_index_map[pid] |
| |
| global GET_WORKER_INDEX |
| GET_WORKER_INDEX = get_worker_index_use_pid |
| |
| |
| def report_test_failure(name, command, output, timeout): |
| global output_lock |
| with output_lock: |
| if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()): |
| print(file=sys.stderr) |
| print(output, file=sys.stderr) |
| if timeout: |
| timeout_str = " (TIMEOUT)" |
| else: |
| timeout_str = "" |
| print("[%s FAILED]%s" % (name, timeout_str), file=sys.stderr) |
| print("Command invoked: %s" % ' '.join(command), file=sys.stderr) |
| update_progress(name) |
| |
| |
| def report_test_pass(name, output): |
| global output_lock |
| with output_lock: |
| update_progress(name) |
| |
| |
| def update_progress(test_name=""): |
| global output_lock, test_counter, total_tests, test_name_len |
| with output_lock: |
| counter_len = len(str(total_tests)) |
| if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()): |
| sys.stderr.write( |
| "\r%*d out of %d test suites processed - %-*s" % |
| (counter_len, test_counter.value, total_tests, |
| test_name_len.value, test_name)) |
| if len(test_name) > test_name_len.value: |
| test_name_len.value = len(test_name) |
| test_counter.value += 1 |
| sys.stdout.flush() |
| sys.stderr.flush() |
| |
| |
| def parse_test_results(output): |
| passes = 0 |
| failures = 0 |
| unexpected_successes = 0 |
| for result in output: |
| pass_count = re.search("^RESULT:.*([0-9]+) passes", |
| result, re.MULTILINE) |
| fail_count = re.search("^RESULT:.*([0-9]+) failures", |
| result, re.MULTILINE) |
| error_count = re.search("^RESULT:.*([0-9]+) errors", |
| result, re.MULTILINE) |
| unexpected_success_count = re.search( |
| "^RESULT:.*([0-9]+) unexpected successes", result, re.MULTILINE) |
| if pass_count is not None: |
| passes = passes + int(pass_count.group(1)) |
| if fail_count is not None: |
| failures = failures + int(fail_count.group(1)) |
| if unexpected_success_count is not None: |
| unexpected_successes = unexpected_successes + \ |
| int(unexpected_success_count.group(1)) |
| if error_count is not None: |
| failures = failures + int(error_count.group(1)) |
| return passes, failures, unexpected_successes |
| |
| |
| class DoTestProcessDriver(process_control.ProcessDriver): |
| """Drives the dotest.py inferior process and handles bookkeeping.""" |
| |
| def __init__(self, output_file, output_file_lock, pid_events, file_name, |
| soft_terminate_timeout): |
| super(DoTestProcessDriver, self).__init__( |
| soft_terminate_timeout=soft_terminate_timeout) |
| self.output_file = output_file |
| self.output_lock = optional_with.optional_with(output_file_lock) |
| self.pid_events = pid_events |
| self.results = None |
| self.file_name = file_name |
| |
| def write(self, content): |
| with self.output_lock: |
| self.output_file.write(content) |
| |
| def on_process_started(self): |
| if self.pid_events: |
| self.pid_events.put_nowait(('created', self.process.pid)) |
| |
| def on_process_exited(self, command, output, was_timeout, exit_status): |
| if self.pid_events: |
| # No point in culling out those with no exit_status (i.e. |
| # those we failed to kill). That would just cause |
| # downstream code to try to kill it later on a Ctrl-C. At |
| # this point, a best-effort-to-kill already took place. So |
| # call it destroyed here. |
| self.pid_events.put_nowait(('destroyed', self.process.pid)) |
| |
| # Override the exit status if it was a timeout. |
| if was_timeout: |
| exit_status = eTimedOut |
| |
| # If we didn't end up with any output, call it empty for |
| # stdout/stderr. |
| if output is None: |
| output = ('', '') |
| |
| # Now parse the output. |
| passes, failures, unexpected_successes = parse_test_results(output) |
| if exit_status == 0: |
| # stdout does not have any useful information from 'dotest.py', |
| # only stderr does. |
| report_test_pass(self.file_name, output[1]) |
| else: |
| report_test_failure( |
| self.file_name, |
| command, |
| output[1], |
| was_timeout) |
| |
| # Save off the results for the caller. |
| self.results = ( |
| self.file_name, |
| exit_status, |
| passes, |
| failures, |
| unexpected_successes) |
| |
| def on_timeout_pre_kill(self): |
| # We're just about to have a timeout take effect. Here's our chance |
| # to do a pre-kill action. |
| |
| # For now, we look to see if the lldbsuite.pre_kill module has a |
| # runner for our platform. |
| module_name = "lldbsuite.pre_kill_hook." + platform.system().lower() |
| import importlib |
| try: |
| module = importlib.import_module(module_name) |
| except ImportError: |
| # We don't have one for this platform. Skip. |
| sys.stderr.write("\nwarning: no timeout handler module: " + |
| module_name + "\n") |
| return |
| |
| # Try to run the pre-kill-hook method. |
| try: |
| # Run the pre-kill command. |
| output_io = StringIO() |
| module.do_pre_kill(self.pid, g_runner_context, output_io) |
| |
| # Write the output to a filename associated with the test file and |
| # pid. |
| MAX_UNCOMPRESSED_BYTE_COUNT = 10 * 1024 |
| |
| content = output_io.getvalue() |
| compress_output = len(content) > MAX_UNCOMPRESSED_BYTE_COUNT |
| basename = "{}-{}.sample".format(self.file_name, self.pid) |
| sample_path = os.path.join(g_session_dir, basename) |
| |
| if compress_output: |
| # Write compressed output into a .zip file. |
| from zipfile import ZipFile, ZIP_DEFLATED |
| zipfile = sample_path + ".zip" |
| with ZipFile(zipfile, "w", ZIP_DEFLATED) as sample_zip: |
| sample_zip.writestr(basename, content) |
| else: |
| # Write raw output into a text file. |
| with open(sample_path, "w") as output_file: |
| output_file.write(content) |
| except Exception as e: |
| sys.stderr.write("caught exception while running " |
| "pre-kill action: {}\n".format(e)) |
| return |
| |
| def is_exceptional_exit(self): |
| """Returns whether the process returned a timeout. |
| |
| Not valid to call until after on_process_exited() completes. |
| |
| @return True if the exit is an exceptional exit (e.g. signal on |
| POSIX); False otherwise. |
| """ |
| if self.results is None: |
| raise Exception( |
| "exit status checked before results are available") |
| return self.process_helper.is_exceptional_exit( |
| self.results[1]) |
| |
| def exceptional_exit_details(self): |
| if self.results is None: |
| raise Exception( |
| "exit status checked before results are available") |
| return self.process_helper.exceptional_exit_details(self.results[1]) |
| |
| def is_timeout(self): |
| if self.results is None: |
| raise Exception( |
| "exit status checked before results are available") |
| return self.results[1] == eTimedOut |
| |
| |
| def get_soft_terminate_timeout(): |
| # Defaults to 10 seconds, but can set |
| # LLDB_TEST_SOFT_TERMINATE_TIMEOUT to a floating point |
| # number in seconds. This value indicates how long |
| # the test runner will wait for the dotest inferior to |
| # handle a timeout via a soft terminate before it will |
| # assume that failed and do a hard terminate. |
| |
| # TODO plumb through command-line option |
| return float(os.environ.get('LLDB_TEST_SOFT_TERMINATE_TIMEOUT', 10.0)) |
| |
| |
| def want_core_on_soft_terminate(): |
| # TODO plumb through command-line option |
| if platform.system() == 'Linux': |
| return True |
| else: |
| return False |
| |
| |
| def send_events_to_collector(events, command): |
| """Sends the given events to the collector described in the command line. |
| |
| @param events the list of events to send to the test event collector. |
| @param command the inferior command line which contains the details on |
| how to connect to the test event collector. |
| """ |
| if events is None or len(events) == 0: |
| # Nothing to do. |
| return |
| |
| # Find the port we need to connect to from the --results-port option. |
| try: |
| arg_index = command.index("--results-port") + 1 |
| except ValueError: |
| # There is no results port, so no way to communicate back to |
| # the event collector. This is not a problem if we're not |
| # using event aggregation. |
| # TODO flag as error once we always use the event system |
| print( |
| "INFO: no event collector, skipping post-inferior test " |
| "event reporting") |
| return |
| |
| if arg_index >= len(command): |
| raise Exception( |
| "expected collector port at index {} in {}".format( |
| arg_index, command)) |
| event_port = int(command[arg_index]) |
| |
| # Create results formatter connected back to collector via socket. |
| config = formatter.FormatterConfig() |
| config.port = event_port |
| formatter_spec = formatter.create_results_formatter(config) |
| if formatter_spec is None or formatter_spec.formatter is None: |
| raise Exception( |
| "Failed to create socket-based ResultsFormatter " |
| "back to test event collector") |
| |
| # Send the events: the port-based event just pickles the content |
| # and sends over to the server side of the socket. |
| for event in events: |
| formatter_spec.formatter.handle_event(event) |
| |
| # Cleanup |
| if formatter_spec.cleanup_func is not None: |
| formatter_spec.cleanup_func() |
| |
| |
| def send_inferior_post_run_events( |
| command, worker_index, process_driver, test_filename): |
| """Sends any test events that should be generated after the inferior runs. |
| |
| These events would include timeouts and exceptional (i.e. signal-returning) |
| process completion results. |
| |
| @param command the list of command parameters passed to subprocess.Popen(). |
| @param worker_index the worker index (possibly None) used to run |
| this process |
| @param process_driver the ProcessDriver-derived instance that was used |
| to run the inferior process. |
| @param test_filename the full path to the Python test file that is being |
| run. |
| """ |
| if process_driver is None: |
| raise Exception("process_driver must not be None") |
| if process_driver.results is None: |
| # Invalid condition - the results should have been set one way or |
| # another, even in a timeout. |
| raise Exception("process_driver.results were not set") |
| |
| # The code below fills in the post events struct. If there are any post |
| # events to fire up, we'll try to make a connection to the socket and |
| # provide the results. |
| post_events = [] |
| |
| # Handle signal/exceptional exits. |
| if process_driver.is_exceptional_exit(): |
| (code, desc) = process_driver.exceptional_exit_details() |
| post_events.append( |
| EventBuilder.event_for_job_exceptional_exit( |
| process_driver.pid, |
| worker_index, |
| code, |
| desc, |
| test_filename, |
| command)) |
| |
| # Handle timeouts. |
| if process_driver.is_timeout(): |
| post_events.append(EventBuilder.event_for_job_timeout( |
| process_driver.pid, |
| worker_index, |
| test_filename, |
| command)) |
| |
| if len(post_events) > 0: |
| send_events_to_collector(post_events, command) |
| |
| |
| def call_with_timeout( |
| command, timeout, name, inferior_pid_events, test_filename): |
| # Add our worker index (if we have one) to all test events |
| # from this inferior. |
| worker_index = None |
| if GET_WORKER_INDEX is not None: |
| try: |
| worker_index = GET_WORKER_INDEX() |
| command.extend([ |
| "--event-add-entries", |
| "worker_index={}:int".format(worker_index)]) |
| except: # pylint: disable=bare-except |
| # Ctrl-C does bad things to multiprocessing.Manager.dict() |
| # lookup. Just swallow it. |
| pass |
| |
| # Create the inferior dotest.py ProcessDriver. |
| soft_terminate_timeout = get_soft_terminate_timeout() |
| want_core = want_core_on_soft_terminate() |
| |
| process_driver = DoTestProcessDriver( |
| sys.stdout, |
| output_lock, |
| inferior_pid_events, |
| name, |
| soft_terminate_timeout) |
| |
| # Run it with a timeout. |
| process_driver.run_command_with_timeout(command, timeout, want_core) |
| |
| # Return the results. |
| if not process_driver.results: |
| # This is truly exceptional. Even a failing or timed out |
| # binary should have called the results-generation code. |
| raise Exception("no test results were generated whatsoever") |
| |
| # Handle cases where the test inferior cannot adequately provide |
| # meaningful results to the test event system. |
| send_inferior_post_run_events( |
| command, |
| worker_index, |
| process_driver, |
| test_filename) |
| |
| return process_driver.results |
| |
| |
| def process_file(test_file, dotest_argv, inferior_pid_events): |
| """Run tests in the specified file in a subprocess and gather the results.""" |
| results = [] |
| base_name = os.path.basename(test_file) |
| |
| import __main__ as main |
| global dotest_options |
| if not dotest_options.p or re.search(dotest_options.p, base_name): |
| script_file = main.__file__ |
| command = ([sys.executable, script_file] + |
| dotest_argv + |
| ["-S", dotest_options.session_file_format] + |
| ["--inferior", "-p", base_name, os.path.dirname(test_file)]) |
| |
| timeout_name = os.path.basename(os.path.splitext(base_name)[0]).upper() |
| |
| timeout = (os.getenv("LLDB_%s_TIMEOUT" % timeout_name) or |
| getDefaultTimeout(dotest_options.lldb_platform_name)) |
| |
| results.append(call_with_timeout( |
| command, timeout, base_name, inferior_pid_events, test_file)) |
| |
| # result = (name, status, passes, failures, unexpected_successes) |
| timed_out = [name for name, status, _, _, _ in results |
| if status == eTimedOut] |
| passed = [name for name, status, _, _, _ in results |
| if status == ePassed] |
| failed = [name for name, status, _, _, _ in results |
| if status != ePassed] |
| unexpected_passes = [ |
| name for name, _, _, _, unexpected_successes in results |
| if unexpected_successes > 0] |
| |
| pass_count = sum([result[2] for result in results]) |
| fail_count = sum([result[3] for result in results]) |
| |
| return ( |
| timed_out, passed, failed, unexpected_passes, pass_count, fail_count) |
| |
| in_q = None |
| out_q = None |
| |
| |
| def process_dir_worker_multiprocessing( |
| a_output_lock, a_test_counter, a_total_tests, a_test_name_len, |
| a_dotest_options, job_queue, result_queue, inferior_pid_events, |
| worker_index_map): |
| """Worker thread main loop when in multiprocessing mode. |
| Takes one directory specification at a time and works on it.""" |
| |
| # Shut off interrupt handling in the child process. |
| signal.signal(signal.SIGINT, signal.SIG_IGN) |
| if hasattr(signal, 'SIGHUP'): |
| signal.signal(signal.SIGHUP, signal.SIG_IGN) |
| |
| # Setup the global state for the worker process. |
| setup_global_variables( |
| a_output_lock, a_test_counter, a_total_tests, a_test_name_len, |
| a_dotest_options, worker_index_map) |
| |
| # Keep grabbing entries from the queue until done. |
| while not job_queue.empty(): |
| try: |
| job = job_queue.get(block=False) |
| result = process_file(job[0], job[1], job[2], |
| inferior_pid_events) |
| result_queue.put(result) |
| except queue.Empty: |
| # Fine, we're done. |
| pass |
| |
| |
| def process_file_worker_multiprocessing_pool(args): |
| return process_file(*args) |
| |
| |
| def process_file_worker_threading(job_queue, result_queue, inferior_pid_events): |
| """Worker thread main loop when in threading mode. |
| |
| This one supports the hand-rolled pooling support. |
| |
| Takes one directory specification at a time and works on it.""" |
| |
| # Keep grabbing entries from the queue until done. |
| while not job_queue.empty(): |
| try: |
| job = job_queue.get(block=False) |
| result = process_file(job[0], job[1], inferior_pid_events) |
| result_queue.put(result) |
| except queue.Empty: |
| # Fine, we're done. |
| pass |
| |
| |
| def process_file_worker_threading_pool(args): |
| return process_file(*args) |
| |
| |
| def process_file_mapper_inprocess(args): |
| """Map adapter for running the subprocess-based, non-threaded test runner. |
| |
| @param args the process work item tuple |
| @return the test result tuple |
| """ |
| return process_file(*args) |
| |
| |
| def collect_active_pids_from_pid_events(event_queue): |
| """ |
| Returns the set of what should be active inferior pids based on |
| the event stream. |
| |
| @param event_queue a multiprocessing.Queue containing events of the |
| form: |
| ('created', pid) |
| ('destroyed', pid) |
| |
| @return set of inferior dotest.py pids activated but never completed. |
| """ |
| active_pid_set = set() |
| while not event_queue.empty(): |
| pid_event = event_queue.get_nowait() |
| if pid_event[0] == 'created': |
| active_pid_set.add(pid_event[1]) |
| elif pid_event[0] == 'destroyed': |
| active_pid_set.remove(pid_event[1]) |
| return active_pid_set |
| |
| |
| def kill_all_worker_processes(workers, inferior_pid_events): |
| """ |
| Kills all specified worker processes and their process tree. |
| |
| @param workers a list of multiprocess.Process worker objects. |
| @param inferior_pid_events a multiprocess.Queue that contains |
| all inferior create and destroy events. Used to construct |
| the list of child pids still outstanding that need to be killed. |
| """ |
| for worker in workers: |
| worker.terminate() |
| worker.join() |
| |
| # Add all the child test pids created. |
| active_pid_set = collect_active_pids_from_pid_events( |
| inferior_pid_events) |
| for inferior_pid in active_pid_set: |
| print("killing inferior pid {}".format(inferior_pid)) |
| os.kill(inferior_pid, signal.SIGKILL) |
| |
| |
| def kill_all_worker_threads(workers, inferior_pid_events): |
| """ |
| Kills all specified worker threads and their process tree. |
| |
| @param workers a list of multiprocess.Process worker objects. |
| @param inferior_pid_events a multiprocess.Queue that contains |
| all inferior create and destroy events. Used to construct |
| the list of child pids still outstanding that need to be killed. |
| """ |
| |
| # Add all the child test pids created. |
| active_pid_set = collect_active_pids_from_pid_events( |
| inferior_pid_events) |
| for inferior_pid in active_pid_set: |
| print("killing inferior pid {}".format(inferior_pid)) |
| os.kill(inferior_pid, signal.SIGKILL) |
| |
| # We don't have a way to nuke the threads. However, since we killed |
| # all the inferiors, and we drained the job queue, this will be |
| # good enough. Wait cleanly for each worker thread to wrap up. |
| for worker in workers: |
| worker.join() |
| |
| |
| def find_test_files_in_dir_tree(dir_root): |
| """Returns all the test files in the given dir hierarchy. |
| |
| @param dir_root the path to the directory to start scanning |
| for test files. All files in this directory and all its children |
| directory trees will be searched. |
| """ |
| for root, _, files in os.walk(dir_root, topdown=False): |
| def is_test_filename(test_dir, base_filename): |
| """Returns True if the given filename matches the test name format. |
| |
| @param test_dir the directory to check. Should be absolute or |
| relative to current working directory. |
| |
| @param base_filename the base name of the filename to check for a |
| dherence to the python test case filename format. |
| |
| @return True if name matches the python test case filename format. |
| """ |
| # Not interested in symbolically linked files. |
| if os.path.islink(os.path.join(test_dir, base_filename)): |
| return False |
| # Only interested in test files with the "Test*.py" naming pattern. |
| return (base_filename.startswith("Test") and |
| base_filename.endswith(".py")) |
| |
| for f in files: |
| if is_test_filename(root, f): |
| yield os.path.join(root, f) |
| |
| |
| def initialize_global_vars_common(num_threads, test_work_items, session_dir, |
| runner_context): |
| global g_session_dir, g_runner_context, total_tests, test_counter |
| global test_name_len |
| |
| total_tests = len(test_work_items) |
| test_counter = multiprocessing.Value('i', 0) |
| test_name_len = multiprocessing.Value('i', 0) |
| g_session_dir = session_dir |
| g_runner_context = runner_context |
| if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()): |
| print( |
| "Testing: %d test suites, %d thread%s" % |
| (total_tests, |
| num_threads, |
| (num_threads > 1) * |
| "s"), |
| file=sys.stderr) |
| update_progress() |
| |
| |
| def initialize_global_vars_multiprocessing(num_threads, test_work_items, |
| session_dir, runner_context): |
| # Initialize the global state we'll use to communicate with the |
| # rest of the flat module. |
| global output_lock |
| output_lock = multiprocessing.RLock() |
| |
| initialize_global_vars_common(num_threads, test_work_items, session_dir, |
| runner_context) |
| |
| |
| def initialize_global_vars_threading(num_threads, test_work_items, session_dir, |
| runner_context): |
| """Initializes global variables used in threading mode. |
| |
| @param num_threads specifies the number of workers used. |
| |
| @param test_work_items specifies all the work items |
| that will be processed. |
| |
| @param session_dir the session directory where test-run-speciif files are |
| written. |
| |
| @param runner_context a dictionary of platform-related data that is passed |
| to the timeout pre-kill hook. |
| """ |
| # Initialize the global state we'll use to communicate with the |
| # rest of the flat module. |
| global output_lock |
| output_lock = threading.RLock() |
| |
| index_lock = threading.RLock() |
| index_map = {} |
| |
| def get_worker_index_threading(): |
| """Returns a 0-based, thread-unique index for the worker thread.""" |
| thread_id = threading.current_thread().ident |
| with index_lock: |
| if thread_id not in index_map: |
| index_map[thread_id] = len(index_map) |
| return index_map[thread_id] |
| |
| global GET_WORKER_INDEX |
| GET_WORKER_INDEX = get_worker_index_threading |
| |
| initialize_global_vars_common(num_threads, test_work_items, session_dir, |
| runner_context) |
| |
| |
| def ctrl_c_loop(main_op_func, done_func, ctrl_c_handler): |
| """Provides a main loop that is Ctrl-C protected. |
| |
| The main loop calls the main_op_func() repeatedly until done_func() |
| returns true. The ctrl_c_handler() method is called with a single |
| int parameter that contains the number of times the ctrl_c has been |
| hit (starting with 1). The ctrl_c_handler() should mutate whatever |
| it needs to have the done_func() return True as soon as it is desired |
| to exit the loop. |
| """ |
| done = False |
| ctrl_c_count = 0 |
| |
| while not done: |
| try: |
| # See if we're done. Start with done check since it is |
| # the first thing executed after a Ctrl-C handler in the |
| # following loop. |
| done = done_func() |
| if not done: |
| # Run the main op once. |
| main_op_func() |
| |
| except KeyboardInterrupt: |
| ctrl_c_count += 1 |
| ctrl_c_handler(ctrl_c_count) |
| |
| |
| def pump_workers_and_asyncore_map(workers, asyncore_map): |
| """Prunes out completed workers and maintains the asyncore loop. |
| |
| The asyncore loop contains the optional socket listener |
| and handlers. When all workers are complete, this method |
| takes care of stopping the listener. It also runs the |
| asyncore loop for the given async map for 10 iterations. |
| |
| @param workers the list of worker Thread/Process instances. |
| |
| @param asyncore_map the asyncore threading-aware map that |
| indicates which channels are in use and still alive. |
| """ |
| |
| # Check on all the workers, removing them from the workers |
| # list as they complete. |
| dead_workers = [] |
| for worker in workers: |
| # This non-blocking join call is what allows us |
| # to still receive keyboard interrupts. |
| worker.join(0.01) |
| if not worker.is_alive(): |
| dead_workers.append(worker) |
| # Clear out the completed workers |
| for dead_worker in dead_workers: |
| workers.remove(dead_worker) |
| |
| # If there are no more workers and there is a listener, |
| # close the listener. |
| global RESULTS_LISTENER_CHANNEL |
| if len(workers) == 0 and RESULTS_LISTENER_CHANNEL is not None: |
| RESULTS_LISTENER_CHANNEL.close() |
| RESULTS_LISTENER_CHANNEL = None |
| |
| # Pump the asyncore map if it isn't empty. |
| if len(asyncore_map) > 0: |
| asyncore.loop(0.1, False, asyncore_map, 10) |
| |
| |
| def handle_ctrl_c(ctrl_c_count, job_queue, workers, inferior_pid_events, |
| stop_all_inferiors_func): |
| """Performs the appropriate ctrl-c action for non-pool parallel test runners |
| |
| @param ctrl_c_count starting with 1, indicates the number of times ctrl-c |
| has been intercepted. The value is 1 on the first intercept, 2 on the |
| second, etc. |
| |
| @param job_queue a Queue object that contains the work still outstanding |
| (i.e. hasn't been assigned to a worker yet). |
| |
| @param workers list of Thread or Process workers. |
| |
| @param inferior_pid_events specifies a Queue of inferior process |
| construction and destruction events. Used to build the list of inferior |
| processes that should be killed if we get that far. |
| |
| @param stop_all_inferiors_func a callable object that takes the |
| workers and inferior_pid_events parameters (in that order) if a hard |
| stop is to be used on the workers. |
| """ |
| |
| # Print out which Ctrl-C we're handling. |
| key_name = [ |
| "first", |
| "second", |
| "third", |
| "many"] |
| |
| if ctrl_c_count < len(key_name): |
| name_index = ctrl_c_count - 1 |
| else: |
| name_index = len(key_name) - 1 |
| message = "\nHandling {} KeyboardInterrupt".format(key_name[name_index]) |
| with output_lock: |
| print(message) |
| |
| if ctrl_c_count == 1: |
| # Remove all outstanding items from the work queue so we stop |
| # doing any more new work. |
| while not job_queue.empty(): |
| try: |
| # Just drain it to stop more work from being started. |
| job_queue.get_nowait() |
| except queue.Empty: |
| pass |
| with output_lock: |
| print("Stopped more work from being started.") |
| elif ctrl_c_count == 2: |
| # Try to stop all inferiors, even the ones currently doing work. |
| stop_all_inferiors_func(workers, inferior_pid_events) |
| else: |
| with output_lock: |
| print("All teardown activities kicked off, should finish soon.") |
| |
| |
| def workers_and_async_done(workers, async_map): |
| """Returns True if the workers list and asyncore channels are all done. |
| |
| @param workers list of workers (threads/processes). These must adhere |
| to the threading Thread or multiprocessing.Process interface. |
| |
| @param async_map the threading-aware asyncore channel map to check |
| for live channels. |
| |
| @return False if the workers list exists and has any entries in it, or |
| if the async_map exists and has any entries left in it; otherwise, True. |
| """ |
| if workers is not None and len(workers) > 0: |
| # We're not done if we still have workers left. |
| return False |
| if async_map is not None and len(async_map) > 0: |
| return False |
| # We're done. |
| return True |
| |
| |
| def multiprocessing_test_runner(num_threads, test_work_items, session_dir, |
| runner_context): |
| """Provides hand-wrapped pooling test runner adapter with Ctrl-C support. |
| |
| This concurrent test runner is based on the multiprocessing |
| library, and rolls its own worker pooling strategy so it |
| can handle Ctrl-C properly. |
| |
| This test runner is known to have an issue running on |
| Windows platforms. |
| |
| @param num_threads the number of worker processes to use. |
| |
| @param test_work_items the iterable of test work item tuples |
| to run. |
| |
| @param session_dir the session directory where test-run-speciif files are |
| written. |
| |
| @param runner_context a dictionary of platform-related data that is passed |
| to the timeout pre-kill hook. |
| """ |
| |
| # Initialize our global state. |
| initialize_global_vars_multiprocessing(num_threads, test_work_items, |
| session_dir, runner_context) |
| |
| # Create jobs. |
| job_queue = multiprocessing.Queue(len(test_work_items)) |
| for test_work_item in test_work_items: |
| job_queue.put(test_work_item) |
| |
| result_queue = multiprocessing.Queue(len(test_work_items)) |
| |
| # Create queues for started child pids. Terminating |
| # the multiprocess processes does not terminate the |
| # child processes they spawn. We can remove this tracking |
| # if/when we move to having the multiprocess process directly |
| # perform the test logic. The Queue size needs to be able to |
| # hold 2 * (num inferior dotest.py processes started) entries. |
| inferior_pid_events = multiprocessing.Queue(4096) |
| |
| # Worker dictionary allows each worker to figure out its worker index. |
| manager = multiprocessing.Manager() |
| worker_index_map = manager.dict() |
| |
| # Create workers. We don't use multiprocessing.Pool due to |
| # challenges with handling ^C keyboard interrupts. |
| workers = [] |
| for _ in range(num_threads): |
| worker = multiprocessing.Process( |
| target=process_file_worker_multiprocessing, |
| args=(output_lock, |
| test_counter, |
| total_tests, |
| test_name_len, |
| dotest_options, |
| job_queue, |
| result_queue, |
| inferior_pid_events, |
| worker_index_map)) |
| worker.start() |
| workers.append(worker) |
| |
| # Main loop: wait for all workers to finish and wait for |
| # the socket handlers to wrap up. |
| ctrl_c_loop( |
| # Main operation of loop |
| lambda: pump_workers_and_asyncore_map( |
| workers, RUNNER_PROCESS_ASYNC_MAP), |
| |
| # Return True when we're done with the main loop. |
| lambda: workers_and_async_done(workers, RUNNER_PROCESS_ASYNC_MAP), |
| |
| # Indicate what we do when we receive one or more Ctrl-Cs. |
| lambda ctrl_c_count: handle_ctrl_c( |
| ctrl_c_count, job_queue, workers, inferior_pid_events, |
| kill_all_worker_processes)) |
| |
| # Reap the test results. |
| test_results = [] |
| while not result_queue.empty(): |
| test_results.append(result_queue.get(block=False)) |
| return test_results |
| |
| |
| def map_async_run_loop(future, channel_map, listener_channel): |
| """Blocks until the Pool.map_async completes and the channel completes. |
| |
| @param future an AsyncResult instance from a Pool.map_async() call. |
| |
| @param channel_map the asyncore dispatch channel map that should be pumped. |
| Optional: may be None. |
| |
| @param listener_channel the channel representing a listener that should be |
| closed once the map_async results are available. |
| |
| @return the results from the async_result instance. |
| """ |
| map_results = None |
| |
| done = False |
| while not done: |
| # Check if we need to reap the map results. |
| if map_results is None: |
| if future.ready(): |
| # Get the results. |
| map_results = future.get() |
| |
| # Close the runner process listener channel if we have |
| # one: no more connections will be incoming. |
| if listener_channel is not None: |
| listener_channel.close() |
| |
| # Pump the asyncore loop if we have a listener socket. |
| if channel_map is not None: |
| asyncore.loop(0.01, False, channel_map, 10) |
| |
| # Figure out if we're done running. |
| done = map_results is not None |
| if channel_map is not None: |
| # We have a runner process async map. Check if it |
| # is complete. |
| if len(channel_map) > 0: |
| # We still have an asyncore channel running. Not done yet. |
| done = False |
| |
| return map_results |
| |
| |
| def multiprocessing_test_runner_pool(num_threads, test_work_items, session_dir, |
| runner_context): |
| # Initialize our global state. |
| initialize_global_vars_multiprocessing(num_threads, test_work_items, |
| session_dir, runner_context) |
| |
| manager = multiprocessing.Manager() |
| worker_index_map = manager.dict() |
| |
| pool = multiprocessing.Pool( |
| num_threads, |
| initializer=setup_global_variables, |
| initargs=(output_lock, test_counter, total_tests, test_name_len, |
| dotest_options, worker_index_map)) |
| |
| # Start the map operation (async mode). |
| map_future = pool.map_async( |
| process_file_worker_multiprocessing_pool, test_work_items) |
| return map_async_run_loop( |
| map_future, RUNNER_PROCESS_ASYNC_MAP, RESULTS_LISTENER_CHANNEL) |
| |
| |
| def threading_test_runner(num_threads, test_work_items, session_dir, |
| runner_context): |
| """Provides hand-wrapped pooling threading-based test runner adapter |
| with Ctrl-C support. |
| |
| This concurrent test runner is based on the threading |
| library, and rolls its own worker pooling strategy so it |
| can handle Ctrl-C properly. |
| |
| @param num_threads the number of worker processes to use. |
| |
| @param test_work_items the iterable of test work item tuples |
| to run. |
| |
| @param session_dir the session directory where test-run-speciif files are |
| written. |
| |
| @param runner_context a dictionary of platform-related data that is passed |
| to the timeout pre-kill hook. |
| """ |
| |
| # Initialize our global state. |
| initialize_global_vars_threading(num_threads, test_work_items, session_dir, |
| runner_context) |
| |
| # Create jobs. |
| job_queue = queue.Queue() |
| for test_work_item in test_work_items: |
| job_queue.put(test_work_item) |
| |
| result_queue = queue.Queue() |
| |
| # Create queues for started child pids. Terminating |
| # the threading threads does not terminate the |
| # child processes they spawn. |
| inferior_pid_events = queue.Queue() |
| |
| # Create workers. We don't use multiprocessing.pool.ThreadedPool |
| # due to challenges with handling ^C keyboard interrupts. |
| workers = [] |
| for _ in range(num_threads): |
| worker = threading.Thread( |
| target=process_file_worker_threading, |
| args=(job_queue, |
| result_queue, |
| inferior_pid_events)) |
| worker.start() |
| workers.append(worker) |
| |
| # Main loop: wait for all workers to finish and wait for |
| # the socket handlers to wrap up. |
| ctrl_c_loop( |
| # Main operation of loop |
| lambda: pump_workers_and_asyncore_map( |
| workers, RUNNER_PROCESS_ASYNC_MAP), |
| |
| # Return True when we're done with the main loop. |
| lambda: workers_and_async_done(workers, RUNNER_PROCESS_ASYNC_MAP), |
| |
| # Indicate what we do when we receive one or more Ctrl-Cs. |
| lambda ctrl_c_count: handle_ctrl_c( |
| ctrl_c_count, job_queue, workers, inferior_pid_events, |
| kill_all_worker_threads)) |
| |
| # Reap the test results. |
| test_results = [] |
| while not result_queue.empty(): |
| test_results.append(result_queue.get(block=False)) |
| return test_results |
| |
| |
| def threading_test_runner_pool(num_threads, test_work_items, session_dir, |
| runner_context): |
| # Initialize our global state. |
| initialize_global_vars_threading(num_threads, test_work_items, session_dir, |
| runner_context) |
| |
| pool = multiprocessing.pool.ThreadPool(num_threads) |
| map_future = pool.map_async( |
| process_file_worker_threading_pool, test_work_items) |
| |
| return map_async_run_loop( |
| map_future, RUNNER_PROCESS_ASYNC_MAP, RESULTS_LISTENER_CHANNEL) |
| |
| |
| def asyncore_run_loop(channel_map): |
| try: |
| asyncore.loop(None, False, channel_map) |
| except: |
| # Swallow it, we're seeing: |
| # error: (9, 'Bad file descriptor') |
| # when the listener channel is closed. Shouldn't be the case. |
| pass |
| |
| |
| def inprocess_exec_test_runner(test_work_items, session_dir, runner_context): |
| # Initialize our global state. |
| initialize_global_vars_multiprocessing(1, test_work_items, session_dir, |
| runner_context) |
| |
| # We're always worker index 0 |
| def get_single_worker_index(): |
| return 0 |
| |
| global GET_WORKER_INDEX |
| GET_WORKER_INDEX = get_single_worker_index |
| |
| # Run the listener and related channel maps in a separate thread. |
| # global RUNNER_PROCESS_ASYNC_MAP |
| global RESULTS_LISTENER_CHANNEL |
| if RESULTS_LISTENER_CHANNEL is not None: |
| socket_thread = threading.Thread( |
| target=lambda: asyncore_run_loop(RUNNER_PROCESS_ASYNC_MAP)) |
| socket_thread.start() |
| |
| # Do the work. |
| test_results = list(map(process_file_mapper_inprocess, test_work_items)) |
| |
| # If we have a listener channel, shut it down here. |
| if RESULTS_LISTENER_CHANNEL is not None: |
| # Close down the channel. |
| RESULTS_LISTENER_CHANNEL.close() |
| RESULTS_LISTENER_CHANNEL = None |
| |
| # Wait for the listener and handlers to complete. |
| socket_thread.join() |
| |
| return test_results |
| |
| |
| def walk_and_invoke(test_files, dotest_argv, num_workers, test_runner_func): |
| """Invokes the test runner on each test file specified by test_files. |
| |
| @param test_files a list of (test_file, full_path_to_test_file) |
| @param num_workers the number of worker queues working on these test files |
| @param test_runner_func the test runner configured to run the tests |
| |
| @return a tuple of results from the running of the specified tests, |
| of the form (timed_out, passed, failed, unexpected_successes, pass_count, |
| fail_count) |
| """ |
| # The async_map is important to keep all thread-related asyncore |
| # channels distinct when we call asyncore.loop() later on. |
| global RESULTS_LISTENER_CHANNEL, RUNNER_PROCESS_ASYNC_MAP |
| RUNNER_PROCESS_ASYNC_MAP = {} |
| |
| # If we're outputting side-channel test results, create the socket |
| # listener channel and tell the inferior to send results to the |
| # port on which we'll be listening. |
| if RESULTS_FORMATTER is not None: |
| forwarding_func = RESULTS_FORMATTER.handle_event |
| RESULTS_LISTENER_CHANNEL = ( |
| dotest_channels.UnpicklingForwardingListenerChannel( |
| RUNNER_PROCESS_ASYNC_MAP, "localhost", 0, |
| 2 * num_workers, forwarding_func)) |
| # Set the results port command line arg. Might have been |
| # inserted previous, so first try to replace. |
| listener_port = str(RESULTS_LISTENER_CHANNEL.address[1]) |
| try: |
| port_value_index = dotest_argv.index("--results-port") + 1 |
| dotest_argv[port_value_index] = listener_port |
| except ValueError: |
| # --results-port doesn't exist (yet), add it |
| dotest_argv.append("--results-port") |
| dotest_argv.append(listener_port) |
| |
| # Build the test work items out of the (dir, file_list) entries passed in. |
| test_work_items = [] |
| for test_file in test_files: |
| test_work_items.append((test_file, dotest_argv, None)) |
| |
| # Convert test work items into test results using whatever |
| # was provided as the test run function. |
| test_results = test_runner_func(test_work_items) |
| |
| # Summarize the results and return to caller. |
| timed_out = sum([result[0] for result in test_results], []) |
| passed = sum([result[1] for result in test_results], []) |
| failed = sum([result[2] for result in test_results], []) |
| unexpected_successes = sum([result[3] for result in test_results], []) |
| pass_count = sum([result[4] for result in test_results]) |
| fail_count = sum([result[5] for result in test_results]) |
| |
| return (timed_out, passed, failed, unexpected_successes, pass_count, |
| fail_count) |
| |
| |
| def getExpectedTimeouts(platform_name): |
| # returns a set of test filenames that might timeout |
| # are we running against a remote target? |
| |
| # Figure out the target system for which we're collecting |
| # the set of expected timeout test filenames. |
| if platform_name is None: |
| target = sys.platform |
| else: |
| m = re.search(r'remote-(\w+)', platform_name) |
| if m is not None: |
| target = m.group(1) |
| else: |
| target = platform_name |
| |
| expected_timeout = set() |
| |
| if target.startswith("freebsd"): |
| expected_timeout |= { |
| "TestBreakpointConditions.py", |
| "TestChangeProcessGroup.py", |
| "TestValueObjectRecursion.py", |
| "TestWatchpointConditionAPI.py", |
| } |
| return expected_timeout |
| |
| |
| def getDefaultTimeout(platform_name): |
| if os.getenv("LLDB_TEST_TIMEOUT"): |
| return os.getenv("LLDB_TEST_TIMEOUT") |
| |
| if platform_name is None: |
| platform_name = sys.platform |
| |
| if platform_name.startswith("remote-"): |
| return "10m" |
| elif platform_name == 'darwin': |
| # We are consistently needing more time on a few tests. |
| return "6m" |
| else: |
| return "4m" |
| |
| |
| def touch(fname, times=None): |
| if os.path.exists(fname): |
| os.utime(fname, times) |
| |
| |
| def find(pattern, path): |
| result = [] |
| for root, dirs, files in os.walk(path): |
| for name in files: |
| if fnmatch.fnmatch(name, pattern): |
| result.append(os.path.join(root, name)) |
| return result |
| |
| |
| def get_test_runner_strategies(num_threads, session_dir, runner_context): |
| """Returns the test runner strategies by name in a dictionary. |
| |
| @param num_threads specifies the number of threads/processes |
| that will be used for concurrent test runners. |
| |
| @param session_dir specifies the session dir to use for |
| auxiliary files. |
| |
| @param runner_context a dictionary of details on the architectures and |
| platform used to run the test suite. This is passed along verbatim to |
| the timeout pre-kill handler, allowing that decoupled component to do |
| process inspection in a platform-specific way. |
| |
| @return dictionary with key as test runner strategy name and |
| value set to a callable object that takes the test work item |
| and returns a test result tuple. |
| """ |
| return { |
| # multiprocessing supports ctrl-c and does not use |
| # multiprocessing.Pool. |
| "multiprocessing": |
| (lambda work_items: multiprocessing_test_runner( |
| num_threads, work_items, session_dir, runner_context)), |
| |
| # multiprocessing-pool uses multiprocessing.Pool but |
| # does not support Ctrl-C. |
| "multiprocessing-pool": |
| (lambda work_items: multiprocessing_test_runner_pool( |
| num_threads, work_items, session_dir, runner_context)), |
| |
| # threading uses a hand-rolled worker pool much |
| # like multiprocessing, but instead uses in-process |
| # worker threads. This one supports Ctrl-C. |
| "threading": |
| (lambda work_items: threading_test_runner( |
| num_threads, work_items, session_dir, runner_context)), |
| |
| # threading-pool uses threading for the workers (in-process) |
| # and uses the multiprocessing.pool thread-enabled pool. |
| # This does not properly support Ctrl-C. |
| "threading-pool": |
| (lambda work_items: threading_test_runner_pool( |
| num_threads, work_items, session_dir, runner_context)), |
| |
| # serial uses the subprocess-based, single process |
| # test runner. This provides process isolation but |
| # no concurrent test execution. |
| "serial": |
| (lambda work_items: inprocess_exec_test_runner( |
| work_items, session_dir, runner_context)) |
| } |
| |
| |
| def _remove_option( |
| args, long_option_name, short_option_name, takes_arg): |
| """Removes option and related option arguments from args array. |
| |
| This method removes all short/long options that match the given |
| arguments. |
| |
| @param args the array of command line arguments (in/out) |
| |
| @param long_option_name the full command line representation of the |
| long-form option that will be removed (including '--'). |
| |
| @param short_option_name the short version of the command line option |
| that will be removed (including '-'). |
| |
| @param takes_arg True if the option takes an argument. |
| |
| """ |
| if long_option_name is not None: |
| regex_string = "^" + long_option_name + "=" |
| long_regex = re.compile(regex_string) |
| if short_option_name is not None: |
| # Short options we only match the -X and assume |
| # any arg is one command line argument jammed together. |
| # i.e. -O--abc=1 is a single argument in the args list. |
| # We don't handle -O --abc=1, as argparse doesn't handle |
| # it, either. |
| regex_string = "^" + short_option_name |
| short_regex = re.compile(regex_string) |
| |
| def remove_long_internal(): |
| """Removes one matching long option from args. |
| @returns True if one was found and removed; False otherwise. |
| """ |
| try: |
| index = args.index(long_option_name) |
| # Handle the exact match case. |
| if takes_arg: |
| removal_count = 2 |
| else: |
| removal_count = 1 |
| del args[index:index + removal_count] |
| return True |
| except ValueError: |
| # Thanks to argparse not handling options with known arguments |
| # like other options parsing libraries (see |
| # https://bugs.python.org/issue9334), we need to support the |
| # --results-formatter-options={second-level-arguments} (note |
| # the equal sign to fool the first-level arguments parser into |
| # not treating the second-level arguments as first-level |
| # options). We're certainly at risk of getting this wrong |
| # since now we're forced into the business of trying to figure |
| # out what is an argument (although I think this |
| # implementation will suffice). |
| for index in range(len(args)): |
| match = long_regex.search(args[index]) |
| if match: |
| del args[index] |
| return True |
| return False |
| |
| def remove_short_internal(): |
| """Removes one matching short option from args. |
| @returns True if one was found and removed; False otherwise. |
| """ |
| for index in range(len(args)): |
| match = short_regex.search(args[index]) |
| if match: |
| del args[index] |
| return True |
| return False |
| |
| removal_count = 0 |
| while long_option_name is not None and remove_long_internal(): |
| removal_count += 1 |
| while short_option_name is not None and remove_short_internal(): |
| removal_count += 1 |
| if removal_count == 0: |
| raise Exception( |
| "failed to find at least one of '{}', '{}' in options".format( |
| long_option_name, short_option_name)) |
| |
| |
| def adjust_inferior_options(dotest_argv): |
| """Adjusts the commandline args array for inferiors. |
| |
| This method adjusts the inferior dotest commandline options based |
| on the parallel test runner's options. Some of the inferior options |
| will need to change to properly handle aggregation functionality. |
| """ |
| global dotest_options |
| |
| # If we don't have a session directory, create one. |
| if not dotest_options.s: |
| # no session log directory, we need to add this to prevent |
| # every dotest invocation from creating its own directory |
| import datetime |
| # The windows platforms don't like ':' in the pathname. |
| timestamp_started = (datetime.datetime.now() |
| .strftime("%Y-%m-%d-%H_%M_%S")) |
| dotest_argv.append('-s') |
| dotest_argv.append(timestamp_started) |
| dotest_options.s = timestamp_started |
| |
| # Adjust inferior results formatter options - if the parallel |
| # test runner is collecting into the user-specified test results, |
| # we'll have inferiors spawn with the --results-port option and |
| # strip the original test runner options. |
| if dotest_options.results_file is not None: |
| _remove_option(dotest_argv, "--results-file", None, True) |
| if dotest_options.results_port is not None: |
| _remove_option(dotest_argv, "--results-port", None, True) |
| if dotest_options.results_formatter is not None: |
| _remove_option(dotest_argv, "--results-formatter", None, True) |
| if dotest_options.results_formatter_options is not None: |
| _remove_option(dotest_argv, "--results-formatter-option", "-O", |
| True) |
| |
| # Remove the --curses shortcut if specified. |
| if dotest_options.curses: |
| _remove_option(dotest_argv, "--curses", None, False) |
| |
| # Remove test runner name if present. |
| if dotest_options.test_runner_name is not None: |
| _remove_option(dotest_argv, "--test-runner-name", None, True) |
| |
| |
| def is_darwin_version_lower_than(target_version): |
| """Checks that os is Darwin and version is lower than target_version. |
| |
| @param target_version the StrictVersion indicating the version |
| we're checking against. |
| |
| @return True if the OS is Darwin (OS X) and the version number of |
| the OS is less than target_version; False in all other cases. |
| """ |
| if platform.system() != 'Darwin': |
| # Can't be Darwin lower than a certain version. |
| return False |
| |
| system_version = distutils.version.StrictVersion(platform.mac_ver()[0]) |
| return seven.cmp_(system_version, target_version) < 0 |
| |
| |
| def default_test_runner_name(num_threads): |
| """Returns the default test runner name for the configuration. |
| |
| @param num_threads the number of threads/workers this test runner is |
| supposed to use. |
| |
| @return the test runner name that should be used by default when |
| no test runner was explicitly called out on the command line. |
| """ |
| if num_threads == 1: |
| # Use the serial runner. |
| test_runner_name = "serial" |
| elif os.name == "nt": |
| # On Windows, Python uses CRT with a low limit on the number of open |
| # files. If you have a lot of cores, the threading-pool runner will |
| # often fail because it exceeds that limit. It's not clear what the |
| # right balance is, so until we can investigate it more deeply, |
| # just use the one that works |
| test_runner_name = "multiprocessing-pool" |
| elif is_darwin_version_lower_than( |
| distutils.version.StrictVersion("10.10.0")): |
| # OS X versions before 10.10 appear to have an issue using |
| # the threading test runner. Fall back to multiprocessing. |
| # Supports Ctrl-C. |
| test_runner_name = "multiprocessing" |
| else: |
| # For everyone else, use the ctrl-c-enabled threading support. |
| # Should use fewer system resources than the multprocessing |
| # variant. |
| test_runner_name = "threading" |
| return test_runner_name |
| |
| |
| def rerun_tests(test_subdir, tests_for_rerun, dotest_argv, session_dir, |
| runner_context): |
| # Build the list of test files to rerun. Some future time we'll |
| # enable re-run by test method so we can constrain the rerun set |
| # to just the method(s) that were in issued within a file. |
| |
| # Sort rerun files into subdirectories. |
| print("\nRerunning the following files:") |
| rerun_files = [] |
| for test_filename in tests_for_rerun.keys(): |
| # Print the file we'll be rerunning |
| test_relative_path = os.path.relpath( |
| test_filename, lldbsuite.lldb_test_root) |
| print(" {}".format(test_relative_path)) |
| |
| rerun_files.append(test_filename) |
| |
| # Do not update legacy counts, I am getting rid of |
| # them so no point adding complicated merge logic here. |
| rerun_thread_count = 1 |
| # Force the parallel test runner to choose a multi-worker strategy. |
| rerun_runner_name = default_test_runner_name(rerun_thread_count + 1) |
| print("rerun will use the '{}' test runner strategy".format( |
| rerun_runner_name)) |
| |
| runner_strategies_by_name = get_test_runner_strategies( |
| rerun_thread_count, session_dir, runner_context) |
| rerun_runner_func = runner_strategies_by_name[ |
| rerun_runner_name] |
| if rerun_runner_func is None: |
| raise Exception( |
| "failed to find rerun test runner " |
| "function named '{}'".format(rerun_runner_name)) |
| |
| walk_and_invoke( |
| rerun_files, |
| dotest_argv, |
| rerun_thread_count, |
| rerun_runner_func) |
| print("\nTest rerun complete\n") |
| |
| |
| def main(num_threads, test_subdir, test_runner_name, results_formatter): |
| """Run dotest.py in inferior mode in parallel. |
| |
| @param num_threads the parsed value of the num-threads command line |
| argument. |
| |
| @param test_subdir optionally specifies a subdir to limit testing |
| within. May be None if the entire test tree is to be used. This subdir |
| is assumed to be relative to the lldb/test root of the test hierarchy. |
| |
| @param test_runner_name if specified, contains the test runner |
| name which selects the strategy used to run the isolated and |
| optionally concurrent test runner. Specify None to allow the |
| system to choose the most appropriate test runner given desired |
| thread count and OS type. |
| |
| @param results_formatter if specified, provides the TestResultsFormatter |
| instance that will format and output test result data from the |
| side-channel test results. When specified, inferior dotest calls |
| will send test results side-channel data over a socket to the parallel |
| test runner, which will forward them on to results_formatter. |
| """ |
| |
| # Do not shut down on sighup. |
| if hasattr(signal, 'SIGHUP'): |
| signal.signal(signal.SIGHUP, signal.SIG_IGN) |
| |
| dotest_argv = sys.argv[1:] |
| |
| global RESULTS_FORMATTER |
| RESULTS_FORMATTER = results_formatter |
| |
| # We can't use sys.path[0] to determine the script directory |
| # because it doesn't work under a debugger |
| parser = dotest_args.create_parser() |
| global dotest_options |
| dotest_options = dotest_args.parse_args(parser, dotest_argv) |
| |
| adjust_inferior_options(dotest_argv) |
| |
| session_dir = os.path.join(os.getcwd(), dotest_options.s) |
| |
| # The root directory was specified on the command line |
| test_directory = os.path.dirname(os.path.realpath(__file__)) |
| if test_subdir and len(test_subdir) > 0: |
| test_subdir = os.path.join(test_directory, test_subdir) |
| if not os.path.isdir(test_subdir): |
| print( |
| 'specified test subdirectory {} is not a valid directory\n' |
| .format(test_subdir)) |
| else: |
| test_subdir = test_directory |
| |
| # clean core files in test tree from previous runs (Linux) |
| cores = find('core.*', test_subdir) |
| for core in cores: |
| os.unlink(core) |
| |
| system_info = " ".join(platform.uname()) |
| |
| # Figure out which test files should be enabled for expected |
| # timeout |
| expected_timeout = getExpectedTimeouts(dotest_options.lldb_platform_name) |
| if results_formatter is not None: |
| results_formatter.set_expected_timeouts_by_basename(expected_timeout) |
| |
| # Setup the test runner context. This is a dictionary of information that |
| # will be passed along to the timeout pre-kill handler and allows for loose |
| # coupling of its implementation. |
| runner_context = { |
| "arch": configuration.arch, |
| "platform_name": configuration.lldb_platform_name, |
| "platform_url": configuration.lldb_platform_url, |
| "platform_working_dir": configuration.lldb_platform_working_dir, |
| } |
| |
| # Figure out which testrunner strategy we'll use. |
| runner_strategies_by_name = get_test_runner_strategies( |
| num_threads, session_dir, runner_context) |
| |
| # If the user didn't specify a test runner strategy, determine |
| # the default now based on number of threads and OS type. |
| if not test_runner_name: |
| test_runner_name = default_test_runner_name(num_threads) |
| |
| if test_runner_name not in runner_strategies_by_name: |
| raise Exception( |
| "specified testrunner name '{}' unknown. Valid choices: {}".format( |
| test_runner_name, |
| list(runner_strategies_by_name.keys()))) |
| test_runner_func = runner_strategies_by_name[test_runner_name] |
| |
| # Do the first test run phase. |
| summary_results = walk_and_invoke( |
| find_test_files_in_dir_tree(test_subdir), |
| dotest_argv, |
| num_threads, |
| test_runner_func) |
| |
| (timed_out, passed, failed, unexpected_successes, pass_count, |
| fail_count) = summary_results |
| |
| # Check if we have any tests to rerun as phase 2. |
| if results_formatter is not None: |
| tests_for_rerun = results_formatter.tests_for_rerun |
| results_formatter.tests_for_rerun = {} |
| |
| if tests_for_rerun is not None and len(tests_for_rerun) > 0: |
| rerun_file_count = len(tests_for_rerun) |
| print("\n{} test files marked for rerun\n".format( |
| rerun_file_count)) |
| |
| # Clear errors charged to any of the files of the tests that |
| # we are rerunning. |
| # https://llvm.org/bugs/show_bug.cgi?id=27423 |
| results_formatter.clear_file_level_issues(tests_for_rerun, |
| sys.stdout) |
| |
| # Check if the number of files exceeds the max cutoff. If so, |
| # we skip the rerun step. |
| if rerun_file_count > configuration.rerun_max_file_threshold: |
| print("Skipping rerun: max rerun file threshold ({}) " |
| "exceeded".format( |
| configuration.rerun_max_file_threshold)) |
| else: |
| rerun_tests(test_subdir, tests_for_rerun, dotest_argv, |
| session_dir, runner_context) |
| |
| # The results formatter - if present - is done now. Tell it to |
| # terminate. |
| if results_formatter is not None: |
| results_formatter.send_terminate_as_needed() |
| |
| timed_out = set(timed_out) |
| num_test_files = len(passed) + len(failed) |
| num_test_cases = pass_count + fail_count |
| |
| # move core files into session dir |
| cores = find('core.*', test_subdir) |
| for core in cores: |
| dst = core.replace(test_directory, "")[1:] |
| dst = dst.replace(os.path.sep, "-") |
| os.rename(core, os.path.join(session_dir, dst)) |
| |
| # remove expected timeouts from failures |
| for xtime in expected_timeout: |
| if xtime in timed_out: |
| timed_out.remove(xtime) |
| failed.remove(xtime) |
| result = "ExpectedTimeout" |
| elif xtime in passed: |
| result = "UnexpectedCompletion" |
| else: |
| result = None # failed |
| |
| if result: |
| test_name = os.path.splitext(xtime)[0] |
| touch(os.path.join(session_dir, "{}-{}".format(result, test_name))) |
| |
| # Only run the old summary logic if we don't have a results formatter |
| # that already prints the summary. |
| print_legacy_summary = results_formatter is None |
| if not print_legacy_summary: |
| # Print summary results. Summarized results at the end always |
| # get printed to stdout, even if --results-file specifies a different |
| # file for, say, xUnit output. |
| results_formatter.print_results(sys.stdout) |
| |
| # Figure out exit code by count of test result types. |
| issue_count = 0 |
| for issue_status in EventBuilder.TESTRUN_ERROR_STATUS_VALUES: |
| issue_count += results_formatter.counts_by_test_result_status( |
| issue_status) |
| |
| # Return with appropriate result code |
| if issue_count > 0: |
| sys.exit(1) |
| else: |
| sys.exit(0) |
| else: |
| # Print the legacy test results summary. |
| print() |
| sys.stdout.write("Ran %d test suites" % num_test_files) |
| if num_test_files > 0: |
| sys.stdout.write(" (%d failed) (%f%%)" % ( |
| len(failed), 100.0 * len(failed) / num_test_files)) |
| print() |
| sys.stdout.write("Ran %d test cases" % num_test_cases) |
| if num_test_cases > 0: |
| sys.stdout.write(" (%d failed) (%f%%)" % ( |
| fail_count, 100.0 * fail_count / num_test_cases)) |
| print() |
| exit_code = 0 |
| |
| if len(failed) > 0: |
| failed.sort() |
| print("Failing Tests (%d)" % len(failed)) |
| for f in failed: |
| print("%s: LLDB (suite) :: %s (%s)" % ( |
| "TIMEOUT" if f in timed_out else "FAIL", f, system_info |
| )) |
| exit_code = 1 |
| |
| if len(unexpected_successes) > 0: |
| unexpected_successes.sort() |
| print("\nUnexpected Successes (%d)" % len(unexpected_successes)) |
| for u in unexpected_successes: |
| print( |
| "UNEXPECTED SUCCESS: LLDB (suite) :: %s (%s)" % |
| (u, system_info)) |
| |
| sys.exit(exit_code) |
| |
| if __name__ == '__main__': |
| sys.stderr.write( |
| "error: dosep.py no longer supports being called directly. " |
| "Please call dotest.py directly. The dosep.py-specific arguments " |
| "have been added under the Parallel processing arguments.\n") |
| sys.exit(128) |