summaryrefslogtreecommitdiffstats
path: root/packages/Python/lldbsuite/test/dosep.py
diff options
context:
space:
mode:
Diffstat (limited to 'packages/Python/lldbsuite/test/dosep.py')
-rw-r--r--packages/Python/lldbsuite/test/dosep.py1680
1 files changed, 1680 insertions, 0 deletions
diff --git a/packages/Python/lldbsuite/test/dosep.py b/packages/Python/lldbsuite/test/dosep.py
new file mode 100644
index 0000000..51275d5
--- /dev/null
+++ b/packages/Python/lldbsuite/test/dosep.py
@@ -0,0 +1,1680 @@
+"""
+Run the test suite using a separate process for each test file.
+
+Each test will run with a time limit of 10 minutes by default.
+
+Override the default time limit of 10 minutes by setting
+the environment variable LLDB_TEST_TIMEOUT.
+
+E.g., export LLDB_TEST_TIMEOUT=10m
+
+Override the time limit for individual tests by setting
+the environment variable LLDB_[TEST NAME]_TIMEOUT.
+
+E.g., export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=2m
+
+Set to "0" to run without time limit.
+
+E.g., export LLDB_TEST_TIMEOUT=0
+or export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=0
+
+To collect core files for timed out tests,
+do the following before running dosep.py
+
+OSX
+ulimit -c unlimited
+sudo sysctl -w kern.corefile=core.%P
+
+Linux:
+ulimit -c unlimited
+echo core.%p | sudo tee /proc/sys/kernel/core_pattern
+"""
+
+from __future__ import print_function
+from __future__ import absolute_import
+
+# system packages and modules
+import asyncore
+import distutils.version
+import fnmatch
+import multiprocessing
+import multiprocessing.pool
+import os
+import platform
+import re
+import signal
+import sys
+import threading
+
+from six.moves import queue
+
+# Our packages and modules
+import lldbsuite
+import lldbsuite.support.seven as seven
+
+from . import configuration
+from . import dotest_channels
+from . import dotest_args
+from . import result_formatter
+
+from .result_formatter import EventBuilder
+
+
+# Todo: Convert this folder layout to be relative-import friendly and
+# don't hack up sys.path like this
+sys.path.append(os.path.join(os.path.dirname(__file__), "test_runner", "lib"))
+import lldb_utils
+import process_control
+
+# Status codes for running command with timeout.
+eTimedOut, ePassed, eFailed = 124, 0, 1
+
+output_lock = None
+test_counter = None
+total_tests = None
+test_name_len = None
+dotest_options = None
+RESULTS_FORMATTER = None
+RUNNER_PROCESS_ASYNC_MAP = None
+RESULTS_LISTENER_CHANNEL = None
+
+"""Contains an optional function pointer that can return the worker index
+ for the given thread/process calling it. Returns a 0-based index."""
+GET_WORKER_INDEX = None
+
+
+def setup_global_variables(
+ lock, counter, total, name_len, options, worker_index_map):
+ global output_lock, test_counter, total_tests, test_name_len
+ global dotest_options
+ output_lock = lock
+ test_counter = counter
+ total_tests = total
+ test_name_len = name_len
+ dotest_options = options
+
+ if worker_index_map is not None:
+ # We'll use the output lock for this to avoid sharing another lock.
+ # This won't be used much.
+ index_lock = lock
+
+ def get_worker_index_use_pid():
+ """Returns a 0-based, process-unique index for the worker."""
+ pid = os.getpid()
+ with index_lock:
+ if pid not in worker_index_map:
+ worker_index_map[pid] = len(worker_index_map)
+ return worker_index_map[pid]
+
+ global GET_WORKER_INDEX
+ GET_WORKER_INDEX = get_worker_index_use_pid
+
+def report_test_failure(name, command, output):
+ global output_lock
+ with output_lock:
+ if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()):
+ print(file=sys.stderr)
+ print(output, file=sys.stderr)
+ print("[%s FAILED]" % name, file=sys.stderr)
+ print("Command invoked: %s" % ' '.join(command), file=sys.stderr)
+ update_progress(name)
+
+
+def report_test_pass(name, output):
+ global output_lock
+ with output_lock:
+ update_progress(name)
+
+
+def update_progress(test_name=""):
+ global output_lock, test_counter, total_tests, test_name_len
+ with output_lock:
+ counter_len = len(str(total_tests))
+ if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()):
+ sys.stderr.write(
+ "\r%*d out of %d test suites processed - %-*s" %
+ (counter_len, test_counter.value, total_tests,
+ test_name_len.value, test_name))
+ if len(test_name) > test_name_len.value:
+ test_name_len.value = len(test_name)
+ test_counter.value += 1
+ sys.stdout.flush()
+ sys.stderr.flush()
+
+
+def parse_test_results(output):
+ passes = 0
+ failures = 0
+ unexpected_successes = 0
+ for result in output:
+ pass_count = re.search("^RESULT:.*([0-9]+) passes",
+ result, re.MULTILINE)
+ fail_count = re.search("^RESULT:.*([0-9]+) failures",
+ result, re.MULTILINE)
+ error_count = re.search("^RESULT:.*([0-9]+) errors",
+ result, re.MULTILINE)
+ unexpected_success_count = re.search("^RESULT:.*([0-9]+) unexpected successes",
+ result, re.MULTILINE)
+ if pass_count is not None:
+ passes = passes + int(pass_count.group(1))
+ if fail_count is not None:
+ failures = failures + int(fail_count.group(1))
+ if unexpected_success_count is not None:
+ unexpected_successes = unexpected_successes + int(unexpected_success_count.group(1))
+ if error_count is not None:
+ failures = failures + int(error_count.group(1))
+ return passes, failures, unexpected_successes
+
+
+class DoTestProcessDriver(process_control.ProcessDriver):
+ """Drives the dotest.py inferior process and handles bookkeeping."""
+ def __init__(self, output_file, output_file_lock, pid_events, file_name,
+ soft_terminate_timeout):
+ super(DoTestProcessDriver, self).__init__(
+ soft_terminate_timeout=soft_terminate_timeout)
+ self.output_file = output_file
+ self.output_lock = lldb_utils.OptionalWith(output_file_lock)
+ self.pid_events = pid_events
+ self.results = None
+ self.file_name = file_name
+
+ def write(self, content):
+ with self.output_lock:
+ self.output_file.write(content)
+
+ def on_process_started(self):
+ if self.pid_events:
+ self.pid_events.put_nowait(('created', self.process.pid))
+
+ def on_process_exited(self, command, output, was_timeout, exit_status):
+ if self.pid_events:
+ # No point in culling out those with no exit_status (i.e.
+ # those we failed to kill). That would just cause
+ # downstream code to try to kill it later on a Ctrl-C. At
+ # this point, a best-effort-to-kill already took place. So
+ # call it destroyed here.
+ self.pid_events.put_nowait(('destroyed', self.process.pid))
+
+ # Override the exit status if it was a timeout.
+ if was_timeout:
+ exit_status = eTimedOut
+
+ # If we didn't end up with any output, call it empty for
+ # stdout/stderr.
+ if output is None:
+ output = ('', '')
+
+ # Now parse the output.
+ passes, failures, unexpected_successes = parse_test_results(output)
+ if exit_status == 0:
+ # stdout does not have any useful information from 'dotest.py',
+ # only stderr does.
+ report_test_pass(self.file_name, output[1])
+ else:
+ report_test_failure(self.file_name, command, output[1])
+
+ # Save off the results for the caller.
+ self.results = (
+ self.file_name,
+ exit_status,
+ passes,
+ failures,
+ unexpected_successes)
+
+ def is_exceptional_exit(self):
+ """Returns whether the process returned a timeout.
+
+ Not valid to call until after on_process_exited() completes.
+
+ @return True if the exit is an exceptional exit (e.g. signal on
+ POSIX); False otherwise.
+ """
+ if self.results is None:
+ raise Exception(
+ "exit status checked before results are available")
+ return self.process_helper.is_exceptional_exit(
+ self.results[1])
+
+ def exceptional_exit_details(self):
+ if self.results is None:
+ raise Exception(
+ "exit status checked before results are available")
+ return self.process_helper.exceptional_exit_details(self.results[1])
+
+ def is_timeout(self):
+ if self.results is None:
+ raise Exception(
+ "exit status checked before results are available")
+ return self.results[1] == eTimedOut
+
+
+def get_soft_terminate_timeout():
+ # Defaults to 10 seconds, but can set
+ # LLDB_TEST_SOFT_TERMINATE_TIMEOUT to a floating point
+ # number in seconds. This value indicates how long
+ # the test runner will wait for the dotest inferior to
+ # handle a timeout via a soft terminate before it will
+ # assume that failed and do a hard terminate.
+
+ # TODO plumb through command-line option
+ return float(os.environ.get('LLDB_TEST_SOFT_TERMINATE_TIMEOUT', 10.0))
+
+
+def want_core_on_soft_terminate():
+ # TODO plumb through command-line option
+ if platform.system() == 'Linux':
+ return True
+ else:
+ return False
+
+
+def send_events_to_collector(events, command):
+ """Sends the given events to the collector described in the command line.
+
+ @param events the list of events to send to the test event collector.
+ @param command the inferior command line which contains the details on
+ how to connect to the test event collector.
+ """
+ if events is None or len(events) == 0:
+ # Nothing to do.
+ return
+
+ # Find the port we need to connect to from the --results-port option.
+ try:
+ arg_index = command.index("--results-port") + 1
+ except ValueError:
+ # There is no results port, so no way to communicate back to
+ # the event collector. This is not a problem if we're not
+ # using event aggregation.
+ # TODO flag as error once we always use the event system
+ print(
+ "INFO: no event collector, skipping post-inferior test "
+ "event reporting")
+ return
+
+ if arg_index >= len(command):
+ raise Exception(
+ "expected collector port at index {} in {}".format(
+ arg_index, command))
+ event_port = int(command[arg_index])
+
+ # Create results formatter connected back to collector via socket.
+ config = result_formatter.FormatterConfig()
+ config.port = event_port
+ formatter_spec = result_formatter.create_results_formatter(config)
+ if formatter_spec is None or formatter_spec.formatter is None:
+ raise Exception(
+ "Failed to create socket-based ResultsFormatter "
+ "back to test event collector")
+
+ # Send the events: the port-based event just pickles the content
+ # and sends over to the server side of the socket.
+ for event in events:
+ formatter_spec.formatter.handle_event(event)
+
+ # Cleanup
+ if formatter_spec.cleanup_func is not None:
+ formatter_spec.cleanup_func()
+
+
+def send_inferior_post_run_events(
+ command, worker_index, process_driver, test_filename):
+ """Sends any test events that should be generated after the inferior runs.
+
+ These events would include timeouts and exceptional (i.e. signal-returning)
+ process completion results.
+
+ @param command the list of command parameters passed to subprocess.Popen().
+ @param worker_index the worker index (possibly None) used to run
+ this process
+ @param process_driver the ProcessDriver-derived instance that was used
+ to run the inferior process.
+ @param test_filename the full path to the Python test file that is being
+ run.
+ """
+ if process_driver is None:
+ raise Exception("process_driver must not be None")
+ if process_driver.results is None:
+ # Invalid condition - the results should have been set one way or
+ # another, even in a timeout.
+ raise Exception("process_driver.results were not set")
+
+ # The code below fills in the post events struct. If there are any post
+ # events to fire up, we'll try to make a connection to the socket and
+ # provide the results.
+ post_events = []
+
+ # Handle signal/exceptional exits.
+ if process_driver.is_exceptional_exit():
+ (code, desc) = process_driver.exceptional_exit_details()
+ post_events.append(
+ EventBuilder.event_for_job_exceptional_exit(
+ process_driver.pid,
+ worker_index,
+ code,
+ desc,
+ test_filename,
+ command))
+
+ # Handle timeouts.
+ if process_driver.is_timeout():
+ post_events.append(EventBuilder.event_for_job_timeout(
+ process_driver.pid,
+ worker_index,
+ test_filename,
+ command))
+
+ if len(post_events) > 0:
+ send_events_to_collector(post_events, command)
+
+
+def call_with_timeout(
+ command, timeout, name, inferior_pid_events, test_filename):
+ # Add our worker index (if we have one) to all test events
+ # from this inferior.
+ worker_index = None
+ if GET_WORKER_INDEX is not None:
+ try:
+ worker_index = GET_WORKER_INDEX()
+ command.extend([
+ "--event-add-entries",
+ "worker_index={}:int".format(worker_index)])
+ except: # pylint: disable=bare-except
+ # Ctrl-C does bad things to multiprocessing.Manager.dict()
+ # lookup. Just swallow it.
+ pass
+
+ # Create the inferior dotest.py ProcessDriver.
+ soft_terminate_timeout = get_soft_terminate_timeout()
+ want_core = want_core_on_soft_terminate()
+
+ process_driver = DoTestProcessDriver(
+ sys.stdout,
+ output_lock,
+ inferior_pid_events,
+ name,
+ soft_terminate_timeout)
+
+ # Run it with a timeout.
+ process_driver.run_command_with_timeout(command, timeout, want_core)
+
+ # Return the results.
+ if not process_driver.results:
+ # This is truly exceptional. Even a failing or timed out
+ # binary should have called the results-generation code.
+ raise Exception("no test results were generated whatsoever")
+
+ # Handle cases where the test inferior cannot adequately provide
+ # meaningful results to the test event system.
+ send_inferior_post_run_events(
+ command,
+ worker_index,
+ process_driver,
+ test_filename)
+
+ return process_driver.results
+
+
+def process_dir(root, files, dotest_argv, inferior_pid_events):
+ """Examine a directory for tests, and invoke any found within it."""
+ results = []
+ for (base_name, full_test_path) in files:
+ import __main__ as main
+ script_file = main.__file__
+ command = ([sys.executable, script_file] +
+ dotest_argv +
+ ["--inferior", "-p", base_name, root])
+
+ timeout_name = os.path.basename(os.path.splitext(base_name)[0]).upper()
+
+ timeout = (os.getenv("LLDB_%s_TIMEOUT" % timeout_name) or
+ getDefaultTimeout(dotest_options.lldb_platform_name))
+
+ results.append(call_with_timeout(
+ command, timeout, base_name, inferior_pid_events, full_test_path))
+
+ # result = (name, status, passes, failures, unexpected_successes)
+ timed_out = [name for name, status, _, _, _ in results
+ if status == eTimedOut]
+ passed = [name for name, status, _, _, _ in results
+ if status == ePassed]
+ failed = [name for name, status, _, _, _ in results
+ if status != ePassed]
+ unexpected_passes = [
+ name for name, _, _, _, unexpected_successes in results
+ if unexpected_successes > 0]
+
+ pass_count = sum([result[2] for result in results])
+ fail_count = sum([result[3] for result in results])
+
+ return (
+ timed_out, passed, failed, unexpected_passes, pass_count, fail_count)
+
+in_q = None
+out_q = None
+
+
+def process_dir_worker_multiprocessing(
+ a_output_lock, a_test_counter, a_total_tests, a_test_name_len,
+ a_dotest_options, job_queue, result_queue, inferior_pid_events,
+ worker_index_map):
+ """Worker thread main loop when in multiprocessing mode.
+ Takes one directory specification at a time and works on it."""
+
+ # Shut off interrupt handling in the child process.
+ signal.signal(signal.SIGINT, signal.SIG_IGN)
+ if hasattr(signal, 'SIGHUP'):
+ signal.signal(signal.SIGHUP, signal.SIG_IGN)
+
+ # Setup the global state for the worker process.
+ setup_global_variables(
+ a_output_lock, a_test_counter, a_total_tests, a_test_name_len,
+ a_dotest_options, worker_index_map)
+
+ # Keep grabbing entries from the queue until done.
+ while not job_queue.empty():
+ try:
+ job = job_queue.get(block=False)
+ result = process_dir(job[0], job[1], job[2],
+ inferior_pid_events)
+ result_queue.put(result)
+ except queue.Empty:
+ # Fine, we're done.
+ pass
+
+
+def process_dir_worker_multiprocessing_pool(args):
+ return process_dir(*args)
+
+
+def process_dir_worker_threading(job_queue, result_queue, inferior_pid_events):
+ """Worker thread main loop when in threading mode.
+
+ This one supports the hand-rolled pooling support.
+
+ Takes one directory specification at a time and works on it."""
+
+ # Keep grabbing entries from the queue until done.
+ while not job_queue.empty():
+ try:
+ job = job_queue.get(block=False)
+ result = process_dir(job[0], job[1], job[2],
+ inferior_pid_events)
+ result_queue.put(result)
+ except queue.Empty:
+ # Fine, we're done.
+ pass
+
+
+def process_dir_worker_threading_pool(args):
+ return process_dir(*args)
+
+
+def process_dir_mapper_inprocess(args):
+ """Map adapter for running the subprocess-based, non-threaded test runner.
+
+ @param args the process work item tuple
+ @return the test result tuple
+ """
+ return process_dir(*args)
+
+
+def collect_active_pids_from_pid_events(event_queue):
+ """
+ Returns the set of what should be active inferior pids based on
+ the event stream.
+
+ @param event_queue a multiprocessing.Queue containing events of the
+ form:
+ ('created', pid)
+ ('destroyed', pid)
+
+ @return set of inferior dotest.py pids activated but never completed.
+ """
+ active_pid_set = set()
+ while not event_queue.empty():
+ pid_event = event_queue.get_nowait()
+ if pid_event[0] == 'created':
+ active_pid_set.add(pid_event[1])
+ elif pid_event[0] == 'destroyed':
+ active_pid_set.remove(pid_event[1])
+ return active_pid_set
+
+
+def kill_all_worker_processes(workers, inferior_pid_events):
+ """
+ Kills all specified worker processes and their process tree.
+
+ @param workers a list of multiprocess.Process worker objects.
+ @param inferior_pid_events a multiprocess.Queue that contains
+ all inferior create and destroy events. Used to construct
+ the list of child pids still outstanding that need to be killed.
+ """
+ for worker in workers:
+ worker.terminate()
+ worker.join()
+
+ # Add all the child test pids created.
+ active_pid_set = collect_active_pids_from_pid_events(
+ inferior_pid_events)
+ for inferior_pid in active_pid_set:
+ print("killing inferior pid {}".format(inferior_pid))
+ os.kill(inferior_pid, signal.SIGKILL)
+
+
+def kill_all_worker_threads(workers, inferior_pid_events):
+ """
+ Kills all specified worker threads and their process tree.
+
+ @param workers a list of multiprocess.Process worker objects.
+ @param inferior_pid_events a multiprocess.Queue that contains
+ all inferior create and destroy events. Used to construct
+ the list of child pids still outstanding that need to be killed.
+ """
+
+ # Add all the child test pids created.
+ active_pid_set = collect_active_pids_from_pid_events(
+ inferior_pid_events)
+ for inferior_pid in active_pid_set:
+ print("killing inferior pid {}".format(inferior_pid))
+ os.kill(inferior_pid, signal.SIGKILL)
+
+ # We don't have a way to nuke the threads. However, since we killed
+ # all the inferiors, and we drained the job queue, this will be
+ # good enough. Wait cleanly for each worker thread to wrap up.
+ for worker in workers:
+ worker.join()
+
+
+def find_test_files_in_dir_tree(dir_root, found_func):
+ """Calls found_func for all the test files in the given dir hierarchy.
+
+ @param dir_root the path to the directory to start scanning
+ for test files. All files in this directory and all its children
+ directory trees will be searched.
+
+ @param found_func a callable object that will be passed
+ the parent directory (relative to dir_root) and the list of
+ test files from within that directory.
+ """
+ for root, _, files in os.walk(dir_root, topdown=False):
+ def is_test_filename(test_dir, base_filename):
+ """Returns True if the given filename matches the test name format.
+
+ @param test_dir the directory to check. Should be absolute or
+ relative to current working directory.
+
+ @param base_filename the base name of the filename to check for a
+ dherence to the python test case filename format.
+
+ @return True if name matches the python test case filename format.
+ """
+ # Not interested in symbolically linked files.
+ if os.path.islink(os.path.join(test_dir, base_filename)):
+ return False
+ # Only interested in test files with the "Test*.py" naming pattern.
+ return (base_filename.startswith("Test") and
+ base_filename.endswith(".py"))
+
+ tests = [
+ (filename, os.path.join(root, filename))
+ for filename in files
+ if is_test_filename(root, filename)]
+ if tests:
+ found_func(root, tests)
+
+
+def initialize_global_vars_common(num_threads, test_work_items):
+ global total_tests, test_counter, test_name_len
+
+ total_tests = sum([len(item[1]) for item in test_work_items])
+ test_counter = multiprocessing.Value('i', 0)
+ test_name_len = multiprocessing.Value('i', 0)
+ if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()):
+ print("Testing: %d test suites, %d thread%s" % (
+ total_tests, num_threads, (num_threads > 1) * "s"), file=sys.stderr)
+ update_progress()
+
+
+def initialize_global_vars_multiprocessing(num_threads, test_work_items):
+ # Initialize the global state we'll use to communicate with the
+ # rest of the flat module.
+ global output_lock
+ output_lock = multiprocessing.RLock()
+
+ initialize_global_vars_common(num_threads, test_work_items)
+
+
+def initialize_global_vars_threading(num_threads, test_work_items):
+ """Initializes global variables used in threading mode.
+ @param num_threads specifies the number of workers used.
+ @param test_work_items specifies all the work items
+ that will be processed.
+ """
+ # Initialize the global state we'll use to communicate with the
+ # rest of the flat module.
+ global output_lock
+ output_lock = threading.RLock()
+
+ index_lock = threading.RLock()
+ index_map = {}
+
+ def get_worker_index_threading():
+ """Returns a 0-based, thread-unique index for the worker thread."""
+ thread_id = threading.current_thread().ident
+ with index_lock:
+ if thread_id not in index_map:
+ index_map[thread_id] = len(index_map)
+ return index_map[thread_id]
+
+
+ global GET_WORKER_INDEX
+ GET_WORKER_INDEX = get_worker_index_threading
+
+ initialize_global_vars_common(num_threads, test_work_items)
+
+
+def ctrl_c_loop(main_op_func, done_func, ctrl_c_handler):
+ """Provides a main loop that is Ctrl-C protected.
+
+ The main loop calls the main_op_func() repeatedly until done_func()
+ returns true. The ctrl_c_handler() method is called with a single
+ int parameter that contains the number of times the ctrl_c has been
+ hit (starting with 1). The ctrl_c_handler() should mutate whatever
+ it needs to have the done_func() return True as soon as it is desired
+ to exit the loop.
+ """
+ done = False
+ ctrl_c_count = 0
+
+ while not done:
+ try:
+ # See if we're done. Start with done check since it is
+ # the first thing executed after a Ctrl-C handler in the
+ # following loop.
+ done = done_func()
+ if not done:
+ # Run the main op once.
+ main_op_func()
+
+ except KeyboardInterrupt:
+ ctrl_c_count += 1
+ ctrl_c_handler(ctrl_c_count)
+
+
+def pump_workers_and_asyncore_map(workers, asyncore_map):
+ """Prunes out completed workers and maintains the asyncore loop.
+
+ The asyncore loop contains the optional socket listener
+ and handlers. When all workers are complete, this method
+ takes care of stopping the listener. It also runs the
+ asyncore loop for the given async map for 10 iterations.
+
+ @param workers the list of worker Thread/Process instances.
+
+ @param asyncore_map the asyncore threading-aware map that
+ indicates which channels are in use and still alive.
+ """
+
+ # Check on all the workers, removing them from the workers
+ # list as they complete.
+ dead_workers = []
+ for worker in workers:
+ # This non-blocking join call is what allows us
+ # to still receive keyboard interrupts.
+ worker.join(0.01)
+ if not worker.is_alive():
+ dead_workers.append(worker)
+ # Clear out the completed workers
+ for dead_worker in dead_workers:
+ workers.remove(dead_worker)
+
+ # If there are no more workers and there is a listener,
+ # close the listener.
+ global RESULTS_LISTENER_CHANNEL
+ if len(workers) == 0 and RESULTS_LISTENER_CHANNEL is not None:
+ RESULTS_LISTENER_CHANNEL.close()
+ RESULTS_LISTENER_CHANNEL = None
+
+ # Pump the asyncore map if it isn't empty.
+ if len(asyncore_map) > 0:
+ asyncore.loop(0.1, False, asyncore_map, 10)
+
+
+def handle_ctrl_c(ctrl_c_count, job_queue, workers, inferior_pid_events,
+ stop_all_inferiors_func):
+ """Performs the appropriate ctrl-c action for non-pool parallel test runners
+
+ @param ctrl_c_count starting with 1, indicates the number of times ctrl-c
+ has been intercepted. The value is 1 on the first intercept, 2 on the
+ second, etc.
+
+ @param job_queue a Queue object that contains the work still outstanding
+ (i.e. hasn't been assigned to a worker yet).
+
+ @param workers list of Thread or Process workers.
+
+ @param inferior_pid_events specifies a Queue of inferior process
+ construction and destruction events. Used to build the list of inferior
+ processes that should be killed if we get that far.
+
+ @param stop_all_inferiors_func a callable object that takes the
+ workers and inferior_pid_events parameters (in that order) if a hard
+ stop is to be used on the workers.
+ """
+
+ # Print out which Ctrl-C we're handling.
+ key_name = [
+ "first",
+ "second",
+ "third",
+ "many"]
+
+ if ctrl_c_count < len(key_name):
+ name_index = ctrl_c_count - 1
+ else:
+ name_index = len(key_name) - 1
+ message = "\nHandling {} KeyboardInterrupt".format(key_name[name_index])
+ with output_lock:
+ print(message)
+
+ if ctrl_c_count == 1:
+ # Remove all outstanding items from the work queue so we stop
+ # doing any more new work.
+ while not job_queue.empty():
+ try:
+ # Just drain it to stop more work from being started.
+ job_queue.get_nowait()
+ except queue.Empty:
+ pass
+ with output_lock:
+ print("Stopped more work from being started.")
+ elif ctrl_c_count == 2:
+ # Try to stop all inferiors, even the ones currently doing work.
+ stop_all_inferiors_func(workers, inferior_pid_events)
+ else:
+ with output_lock:
+ print("All teardown activities kicked off, should finish soon.")
+
+
+def workers_and_async_done(workers, async_map):
+ """Returns True if the workers list and asyncore channels are all done.
+
+ @param workers list of workers (threads/processes). These must adhere
+ to the threading Thread or multiprocessing.Process interface.
+
+ @param async_map the threading-aware asyncore channel map to check
+ for live channels.
+
+ @return False if the workers list exists and has any entries in it, or
+ if the async_map exists and has any entries left in it; otherwise, True.
+ """
+ if workers is not None and len(workers) > 0:
+ # We're not done if we still have workers left.
+ return False
+ if async_map is not None and len(async_map) > 0:
+ return False
+ # We're done.
+ return True
+
+
+def multiprocessing_test_runner(num_threads, test_work_items):
+ """Provides hand-wrapped pooling test runner adapter with Ctrl-C support.
+
+ This concurrent test runner is based on the multiprocessing
+ library, and rolls its own worker pooling strategy so it
+ can handle Ctrl-C properly.
+
+ This test runner is known to have an issue running on
+ Windows platforms.
+
+ @param num_threads the number of worker processes to use.
+
+ @param test_work_items the iterable of test work item tuples
+ to run.
+ """
+
+ # Initialize our global state.
+ initialize_global_vars_multiprocessing(num_threads, test_work_items)
+
+ # Create jobs.
+ job_queue = multiprocessing.Queue(len(test_work_items))
+ for test_work_item in test_work_items:
+ job_queue.put(test_work_item)
+
+ result_queue = multiprocessing.Queue(len(test_work_items))
+
+ # Create queues for started child pids. Terminating
+ # the multiprocess processes does not terminate the
+ # child processes they spawn. We can remove this tracking
+ # if/when we move to having the multiprocess process directly
+ # perform the test logic. The Queue size needs to be able to
+ # hold 2 * (num inferior dotest.py processes started) entries.
+ inferior_pid_events = multiprocessing.Queue(4096)
+
+ # Worker dictionary allows each worker to figure out its worker index.
+ manager = multiprocessing.Manager()
+ worker_index_map = manager.dict()
+
+ # Create workers. We don't use multiprocessing.Pool due to
+ # challenges with handling ^C keyboard interrupts.
+ workers = []
+ for _ in range(num_threads):
+ worker = multiprocessing.Process(
+ target=process_dir_worker_multiprocessing,
+ args=(output_lock,
+ test_counter,
+ total_tests,
+ test_name_len,
+ dotest_options,
+ job_queue,
+ result_queue,
+ inferior_pid_events,
+ worker_index_map))
+ worker.start()
+ workers.append(worker)
+
+ # Main loop: wait for all workers to finish and wait for
+ # the socket handlers to wrap up.
+ ctrl_c_loop(
+ # Main operation of loop
+ lambda: pump_workers_and_asyncore_map(
+ workers, RUNNER_PROCESS_ASYNC_MAP),
+
+ # Return True when we're done with the main loop.
+ lambda: workers_and_async_done(workers, RUNNER_PROCESS_ASYNC_MAP),
+
+ # Indicate what we do when we receive one or more Ctrl-Cs.
+ lambda ctrl_c_count: handle_ctrl_c(
+ ctrl_c_count, job_queue, workers, inferior_pid_events,
+ kill_all_worker_processes))
+
+ # Reap the test results.
+ test_results = []
+ while not result_queue.empty():
+ test_results.append(result_queue.get(block=False))
+ return test_results
+
+
+def map_async_run_loop(future, channel_map, listener_channel):
+ """Blocks until the Pool.map_async completes and the channel completes.
+
+ @param future an AsyncResult instance from a Pool.map_async() call.
+
+ @param channel_map the asyncore dispatch channel map that should be pumped.
+ Optional: may be None.
+
+ @param listener_channel the channel representing a listener that should be
+ closed once the map_async results are available.
+
+ @return the results from the async_result instance.
+ """
+ map_results = None
+
+ done = False
+ while not done:
+ # Check if we need to reap the map results.
+ if map_results is None:
+ if future.ready():
+ # Get the results.
+ map_results = future.get()
+
+ # Close the runner process listener channel if we have
+ # one: no more connections will be incoming.
+ if listener_channel is not None:
+ listener_channel.close()
+
+ # Pump the asyncore loop if we have a listener socket.
+ if channel_map is not None:
+ asyncore.loop(0.01, False, channel_map, 10)
+
+ # Figure out if we're done running.
+ done = map_results is not None
+ if channel_map is not None:
+ # We have a runner process async map. Check if it
+ # is complete.
+ if len(channel_map) > 0:
+ # We still have an asyncore channel running. Not done yet.
+ done = False
+
+ return map_results
+
+
+def multiprocessing_test_runner_pool(num_threads, test_work_items):
+ # Initialize our global state.
+ initialize_global_vars_multiprocessing(num_threads, test_work_items)
+
+ manager = multiprocessing.Manager()
+ worker_index_map = manager.dict()
+
+ pool = multiprocessing.Pool(
+ num_threads,
+ initializer=setup_global_variables,
+ initargs=(output_lock, test_counter, total_tests, test_name_len,
+ dotest_options, worker_index_map))
+
+ # Start the map operation (async mode).
+ map_future = pool.map_async(
+ process_dir_worker_multiprocessing_pool, test_work_items)
+ return map_async_run_loop(
+ map_future, RUNNER_PROCESS_ASYNC_MAP, RESULTS_LISTENER_CHANNEL)
+
+
+def threading_test_runner(num_threads, test_work_items):
+ """Provides hand-wrapped pooling threading-based test runner adapter
+ with Ctrl-C support.
+
+ This concurrent test runner is based on the threading
+ library, and rolls its own worker pooling strategy so it
+ can handle Ctrl-C properly.
+
+ @param num_threads the number of worker processes to use.
+
+ @param test_work_items the iterable of test work item tuples
+ to run.
+ """
+
+ # Initialize our global state.
+ initialize_global_vars_threading(num_threads, test_work_items)
+
+ # Create jobs.
+ job_queue = queue.Queue()
+ for test_work_item in test_work_items:
+ job_queue.put(test_work_item)
+
+ result_queue = queue.Queue()
+
+ # Create queues for started child pids. Terminating
+ # the threading threads does not terminate the
+ # child processes they spawn.
+ inferior_pid_events = queue.Queue()
+
+ # Create workers. We don't use multiprocessing.pool.ThreadedPool
+ # due to challenges with handling ^C keyboard interrupts.
+ workers = []
+ for _ in range(num_threads):
+ worker = threading.Thread(
+ target=process_dir_worker_threading,
+ args=(job_queue,
+ result_queue,
+ inferior_pid_events))
+ worker.start()
+ workers.append(worker)
+
+ # Main loop: wait for all workers to finish and wait for
+ # the socket handlers to wrap up.
+ ctrl_c_loop(
+ # Main operation of loop
+ lambda: pump_workers_and_asyncore_map(
+ workers, RUNNER_PROCESS_ASYNC_MAP),
+
+ # Return True when we're done with the main loop.
+ lambda: workers_and_async_done(workers, RUNNER_PROCESS_ASYNC_MAP),
+
+ # Indicate what we do when we receive one or more Ctrl-Cs.
+ lambda ctrl_c_count: handle_ctrl_c(
+ ctrl_c_count, job_queue, workers, inferior_pid_events,
+ kill_all_worker_threads))
+
+ # Reap the test results.
+ test_results = []
+ while not result_queue.empty():
+ test_results.append(result_queue.get(block=False))
+ return test_results
+
+
+def threading_test_runner_pool(num_threads, test_work_items):
+ # Initialize our global state.
+ initialize_global_vars_threading(num_threads, test_work_items)
+
+ pool = multiprocessing.pool.ThreadPool(num_threads)
+ map_future = pool.map_async(
+ process_dir_worker_threading_pool, test_work_items)
+
+ return map_async_run_loop(
+ map_future, RUNNER_PROCESS_ASYNC_MAP, RESULTS_LISTENER_CHANNEL)
+
+
+def asyncore_run_loop(channel_map):
+ try:
+ asyncore.loop(None, False, channel_map)
+ except:
+ # Swallow it, we're seeing:
+ # error: (9, 'Bad file descriptor')
+ # when the listener channel is closed. Shouldn't be the case.
+ pass
+
+
+def inprocess_exec_test_runner(test_work_items):
+ # Initialize our global state.
+ initialize_global_vars_multiprocessing(1, test_work_items)
+
+ # We're always worker index 0
+ global GET_WORKER_INDEX
+ GET_WORKER_INDEX = lambda: 0
+
+ # Run the listener and related channel maps in a separate thread.
+ # global RUNNER_PROCESS_ASYNC_MAP
+ global RESULTS_LISTENER_CHANNEL
+ if RESULTS_LISTENER_CHANNEL is not None:
+ socket_thread = threading.Thread(
+ target=lambda: asyncore_run_loop(RUNNER_PROCESS_ASYNC_MAP))
+ socket_thread.start()
+
+ # Do the work.
+ test_results = list(map(process_dir_mapper_inprocess, test_work_items))
+
+ # If we have a listener channel, shut it down here.
+ if RESULTS_LISTENER_CHANNEL is not None:
+ # Close down the channel.
+ RESULTS_LISTENER_CHANNEL.close()
+ RESULTS_LISTENER_CHANNEL = None
+
+ # Wait for the listener and handlers to complete.
+ socket_thread.join()
+
+ return test_results
+
+def walk_and_invoke(test_files, dotest_argv, num_workers, test_runner_func):
+ """Invokes the test runner on each test file specified by test_files.
+
+ @param test_files a list of (test_subdir, list_of_test_files_in_dir)
+ @param num_workers the number of worker queues working on these test files
+ @param test_runner_func the test runner configured to run the tests
+
+ @return a tuple of results from the running of the specified tests,
+ of the form (timed_out, passed, failed, unexpected_successes, pass_count,
+ fail_count)
+ """
+ # The async_map is important to keep all thread-related asyncore
+ # channels distinct when we call asyncore.loop() later on.
+ global RESULTS_LISTENER_CHANNEL, RUNNER_PROCESS_ASYNC_MAP
+ RUNNER_PROCESS_ASYNC_MAP = {}
+
+ # If we're outputting side-channel test results, create the socket
+ # listener channel and tell the inferior to send results to the
+ # port on which we'll be listening.
+ if RESULTS_FORMATTER is not None:
+ forwarding_func = RESULTS_FORMATTER.handle_event
+ RESULTS_LISTENER_CHANNEL = (
+ dotest_channels.UnpicklingForwardingListenerChannel(
+ RUNNER_PROCESS_ASYNC_MAP, "localhost", 0,
+ 2 * num_workers, forwarding_func))
+ # Set the results port command line arg. Might have been
+ # inserted previous, so first try to replace.
+ listener_port = str(RESULTS_LISTENER_CHANNEL.address[1])
+ try:
+ port_value_index = dotest_argv.index("--results-port") + 1
+ dotest_argv[port_value_index] = listener_port
+ except ValueError:
+ # --results-port doesn't exist (yet), add it
+ dotest_argv.append("--results-port")
+ dotest_argv.append(listener_port)
+
+ # Build the test work items out of the (dir, file_list) entries passed in.
+ test_work_items = []
+ for entry in test_files:
+ test_work_items.append((entry[0], entry[1], dotest_argv, None))
+
+ # Convert test work items into test results using whatever
+ # was provided as the test run function.
+ test_results = test_runner_func(test_work_items)
+
+ # Summarize the results and return to caller.
+ timed_out = sum([result[0] for result in test_results], [])
+ passed = sum([result[1] for result in test_results], [])
+ failed = sum([result[2] for result in test_results], [])
+ unexpected_successes = sum([result[3] for result in test_results], [])
+ pass_count = sum([result[4] for result in test_results])
+ fail_count = sum([result[5] for result in test_results])
+
+ return (timed_out, passed, failed, unexpected_successes, pass_count,
+ fail_count)
+
+
+def getExpectedTimeouts(platform_name):
+ # returns a set of test filenames that might timeout
+ # are we running against a remote target?
+ host = sys.platform
+ if platform_name is None:
+ target = sys.platform
+ else:
+ m = re.search(r'remote-(\w+)', platform_name)
+ target = m.group(1)
+
+ expected_timeout = set()
+
+ if target.startswith("android"):
+ expected_timeout |= {
+ "TestExitDuringStep.py",
+ "TestHelloWorld.py",
+ }
+ if host.startswith("win32"):
+ expected_timeout |= {
+ "TestEvents.py",
+ "TestThreadStates.py",
+ }
+ elif target.startswith("freebsd"):
+ expected_timeout |= {
+ "TestBreakpointConditions.py",
+ "TestChangeProcessGroup.py",
+ "TestValueObjectRecursion.py",
+ "TestWatchpointConditionAPI.py",
+ }
+ elif target.startswith("darwin"):
+ expected_timeout |= {
+ # times out on MBP Retina, Mid 2012
+ "TestThreadSpecificBreakpoint.py",
+ "TestExitDuringStep.py",
+ "TestIntegerTypesExpr.py",
+ }
+ return expected_timeout
+
+
+def getDefaultTimeout(platform_name):
+ if os.getenv("LLDB_TEST_TIMEOUT"):
+ return os.getenv("LLDB_TEST_TIMEOUT")
+
+ if platform_name is None:
+ platform_name = sys.platform
+
+ if platform_name.startswith("remote-"):
+ return "10m"
+ elif platform_name == 'darwin':
+ # We are consistently needing more time on a few tests.
+ return "6m"
+ else:
+ return "4m"
+
+
+def touch(fname, times=None):
+ if os.path.exists(fname):
+ os.utime(fname, times)
+
+
+def find(pattern, path):
+ result = []
+ for root, dirs, files in os.walk(path):
+ for name in files:
+ if fnmatch.fnmatch(name, pattern):
+ result.append(os.path.join(root, name))
+ return result
+
+
+def get_test_runner_strategies(num_threads):
+ """Returns the test runner strategies by name in a dictionary.
+
+ @param num_threads specifies the number of threads/processes
+ that will be used for concurrent test runners.
+
+ @return dictionary with key as test runner strategy name and
+ value set to a callable object that takes the test work item
+ and returns a test result tuple.
+ """
+ return {
+ # multiprocessing supports ctrl-c and does not use
+ # multiprocessing.Pool.
+ "multiprocessing":
+ (lambda work_items: multiprocessing_test_runner(
+ num_threads, work_items)),
+
+ # multiprocessing-pool uses multiprocessing.Pool but
+ # does not support Ctrl-C.
+ "multiprocessing-pool":
+ (lambda work_items: multiprocessing_test_runner_pool(
+ num_threads, work_items)),
+
+ # threading uses a hand-rolled worker pool much
+ # like multiprocessing, but instead uses in-process
+ # worker threads. This one supports Ctrl-C.
+ "threading":
+ (lambda work_items: threading_test_runner(num_threads, work_items)),
+
+ # threading-pool uses threading for the workers (in-process)
+ # and uses the multiprocessing.pool thread-enabled pool.
+ # This does not properly support Ctrl-C.
+ "threading-pool":
+ (lambda work_items: threading_test_runner_pool(
+ num_threads, work_items)),
+
+ # serial uses the subprocess-based, single process
+ # test runner. This provides process isolation but
+ # no concurrent test execution.
+ "serial":
+ inprocess_exec_test_runner
+ }
+
+
+def _remove_option(
+ args, long_option_name, short_option_name, takes_arg):
+ """Removes option and related option arguments from args array.
+
+ This method removes all short/long options that match the given
+ arguments.
+
+ @param args the array of command line arguments (in/out)
+
+ @param long_option_name the full command line representation of the
+ long-form option that will be removed (including '--').
+
+ @param short_option_name the short version of the command line option
+ that will be removed (including '-').
+
+ @param takes_arg True if the option takes an argument.
+
+ """
+ if long_option_name is not None:
+ regex_string = "^" + long_option_name + "="
+ long_regex = re.compile(regex_string)
+ if short_option_name is not None:
+ # Short options we only match the -X and assume
+ # any arg is one command line argument jammed together.
+ # i.e. -O--abc=1 is a single argument in the args list.
+ # We don't handle -O --abc=1, as argparse doesn't handle
+ # it, either.
+ regex_string = "^" + short_option_name
+ short_regex = re.compile(regex_string)
+
+ def remove_long_internal():
+ """Removes one matching long option from args.
+ @returns True if one was found and removed; False otherwise.
+ """
+ try:
+ index = args.index(long_option_name)
+ # Handle the exact match case.
+ if takes_arg:
+ removal_count = 2
+ else:
+ removal_count = 1
+ del args[index:index+removal_count]
+ return True
+ except ValueError:
+ # Thanks to argparse not handling options with known arguments
+ # like other options parsing libraries (see
+ # https://bugs.python.org/issue9334), we need to support the
+ # --results-formatter-options={second-level-arguments} (note
+ # the equal sign to fool the first-level arguments parser into
+ # not treating the second-level arguments as first-level
+ # options). We're certainly at risk of getting this wrong
+ # since now we're forced into the business of trying to figure
+ # out what is an argument (although I think this
+ # implementation will suffice).
+ for index in range(len(args)):
+ match = long_regex.search(args[index])
+ if match:
+ del args[index]
+ return True
+ return False
+
+ def remove_short_internal():
+ """Removes one matching short option from args.
+ @returns True if one was found and removed; False otherwise.
+ """
+ for index in range(len(args)):
+ match = short_regex.search(args[index])
+ if match:
+ del args[index]
+ return True
+ return False
+
+ removal_count = 0
+ while long_option_name is not None and remove_long_internal():
+ removal_count += 1
+ while short_option_name is not None and remove_short_internal():
+ removal_count += 1
+ if removal_count == 0:
+ raise Exception(
+ "failed to find at least one of '{}', '{}' in options".format(
+ long_option_name, short_option_name))
+
+
+def adjust_inferior_options(dotest_argv):
+ """Adjusts the commandline args array for inferiors.
+
+ This method adjusts the inferior dotest commandline options based
+ on the parallel test runner's options. Some of the inferior options
+ will need to change to properly handle aggregation functionality.
+ """
+ global dotest_options
+
+ # If we don't have a session directory, create one.
+ if not dotest_options.s:
+ # no session log directory, we need to add this to prevent
+ # every dotest invocation from creating its own directory
+ import datetime
+ # The windows platforms don't like ':' in the pathname.
+ timestamp_started = datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S")
+ dotest_argv.append('-s')
+ dotest_argv.append(timestamp_started)
+ dotest_options.s = timestamp_started
+
+ # Adjust inferior results formatter options - if the parallel
+ # test runner is collecting into the user-specified test results,
+ # we'll have inferiors spawn with the --results-port option and
+ # strip the original test runner options.
+ if dotest_options.results_file is not None:
+ _remove_option(dotest_argv, "--results-file", None, True)
+ if dotest_options.results_port is not None:
+ _remove_option(dotest_argv, "--results-port", None, True)
+ if dotest_options.results_formatter is not None:
+ _remove_option(dotest_argv, "--results-formatter", None, True)
+ if dotest_options.results_formatter_options is not None:
+ _remove_option(dotest_argv, "--results-formatter-option", "-O",
+ True)
+
+ # Remove the --curses shortcut if specified.
+ if dotest_options.curses:
+ _remove_option(dotest_argv, "--curses", None, False)
+
+ # Remove test runner name if present.
+ if dotest_options.test_runner_name is not None:
+ _remove_option(dotest_argv, "--test-runner-name", None, True)
+
+
+def is_darwin_version_lower_than(target_version):
+ """Checks that os is Darwin and version is lower than target_version.
+
+ @param target_version the StrictVersion indicating the version
+ we're checking against.
+
+ @return True if the OS is Darwin (OS X) and the version number of
+ the OS is less than target_version; False in all other cases.
+ """
+ if platform.system() != 'Darwin':
+ # Can't be Darwin lower than a certain version.
+ return False
+
+ system_version = distutils.version.StrictVersion(platform.mac_ver()[0])
+ return seven.cmp_(system_version, target_version) < 0
+
+
+def default_test_runner_name(num_threads):
+ """Returns the default test runner name for the configuration.
+
+ @param num_threads the number of threads/workers this test runner is
+ supposed to use.
+
+ @return the test runner name that should be used by default when
+ no test runner was explicitly called out on the command line.
+ """
+ if num_threads == 1:
+ # Use the serial runner.
+ test_runner_name = "serial"
+ elif os.name == "nt":
+ # On Windows, Python uses CRT with a low limit on the number of open
+ # files. If you have a lot of cores, the threading-pool runner will
+ # often fail because it exceeds that limit. It's not clear what the
+ # right balance is, so until we can investigate it more deeply,
+ # just use the one that works
+ test_runner_name = "multiprocessing-pool"
+ elif is_darwin_version_lower_than(
+ distutils.version.StrictVersion("10.10.0")):
+ # OS X versions before 10.10 appear to have an issue using
+ # the threading test runner. Fall back to multiprocessing.
+ # Supports Ctrl-C.
+ test_runner_name = "multiprocessing"
+ else:
+ # For everyone else, use the ctrl-c-enabled threading support.
+ # Should use fewer system resources than the multprocessing
+ # variant.
+ test_runner_name = "threading"
+ return test_runner_name
+
+
+def rerun_tests(test_subdir, tests_for_rerun, dotest_argv):
+ # Build the list of test files to rerun. Some future time we'll
+ # enable re-run by test method so we can constrain the rerun set
+ # to just the method(s) that were in issued within a file.
+
+ # Sort rerun files into subdirectories.
+ print("\nRerunning the following files:")
+ rerun_files_by_subdir = {}
+ for test_filename in tests_for_rerun.keys():
+ # Print the file we'll be rerunning
+ test_relative_path = os.path.relpath(
+ test_filename, lldbsuite.lldb_test_root)
+ print(" {}".format(test_relative_path))
+
+ # Store test filenames by subdir.
+ test_dir = os.path.dirname(test_filename)
+ test_basename = os.path.basename(test_filename)
+ if test_dir in rerun_files_by_subdir:
+ rerun_files_by_subdir[test_dir].append(
+ (test_basename, test_filename))
+ else:
+ rerun_files_by_subdir[test_dir] = [(test_basename, test_filename)]
+
+ # Break rerun work up by subdirectory. We do this since
+ # we have an invariant that states only one test file can
+ # be run at a time in any given subdirectory (related to
+ # rules around built inferior test program lifecycle).
+ rerun_work = []
+ for files_by_subdir in rerun_files_by_subdir.values():
+ rerun_work.append((test_subdir, files_by_subdir))
+
+ # Run the work with the serial runner.
+ # Do not update legacy counts, I am getting rid of
+ # them so no point adding complicated merge logic here.
+ rerun_thread_count = 1
+ # Force the parallel test runner to choose a multi-worker strategy.
+ rerun_runner_name = default_test_runner_name(rerun_thread_count + 1)
+ print("rerun will use the '{}' test runner strategy".format(
+ rerun_runner_name))
+
+ runner_strategies_by_name = get_test_runner_strategies(rerun_thread_count)
+ rerun_runner_func = runner_strategies_by_name[
+ rerun_runner_name]
+ if rerun_runner_func is None:
+ raise Exception(
+ "failed to find rerun test runner "
+ "function named '{}'".format(rerun_runner_name))
+
+ walk_and_invoke(
+ rerun_work,
+ dotest_argv,
+ rerun_thread_count,
+ rerun_runner_func)
+ print("\nTest rerun complete\n")
+
+
+def main(num_threads, test_subdir, test_runner_name, results_formatter):
+ """Run dotest.py in inferior mode in parallel.
+
+ @param num_threads the parsed value of the num-threads command line
+ argument.
+
+ @param test_subdir optionally specifies a subdir to limit testing
+ within. May be None if the entire test tree is to be used. This subdir
+ is assumed to be relative to the lldb/test root of the test hierarchy.
+
+ @param test_runner_name if specified, contains the test runner
+ name which selects the strategy used to run the isolated and
+ optionally concurrent test runner. Specify None to allow the
+ system to choose the most appropriate test runner given desired
+ thread count and OS type.
+
+ @param results_formatter if specified, provides the TestResultsFormatter
+ instance that will format and output test result data from the
+ side-channel test results. When specified, inferior dotest calls
+ will send test results side-channel data over a socket to the parallel
+ test runner, which will forward them on to results_formatter.
+ """
+
+ # Do not shut down on sighup.
+ if hasattr(signal, 'SIGHUP'):
+ signal.signal(signal.SIGHUP, signal.SIG_IGN)
+
+ dotest_argv = sys.argv[1:]
+
+ global RESULTS_FORMATTER
+ RESULTS_FORMATTER = results_formatter
+
+ # We can't use sys.path[0] to determine the script directory
+ # because it doesn't work under a debugger
+ parser = dotest_args.create_parser()
+ global dotest_options
+ dotest_options = dotest_args.parse_args(parser, dotest_argv)
+
+ adjust_inferior_options(dotest_argv)
+
+ session_dir = os.path.join(os.getcwd(), dotest_options.s)
+
+ # The root directory was specified on the command line
+ test_directory = os.path.dirname(os.path.realpath(__file__))
+ if test_subdir and len(test_subdir) > 0:
+ test_subdir = os.path.join(test_directory, test_subdir)
+ else:
+ test_subdir = test_directory
+
+ # clean core files in test tree from previous runs (Linux)
+ cores = find('core.*', test_subdir)
+ for core in cores:
+ os.unlink(core)
+
+ system_info = " ".join(platform.uname())
+
+ # Figure out which test files should be enabled for expected
+ # timeout
+ expected_timeout = getExpectedTimeouts(dotest_options.lldb_platform_name)
+ if results_formatter is not None:
+ results_formatter.set_expected_timeouts_by_basename(expected_timeout)
+
+ # Figure out which testrunner strategy we'll use.
+ runner_strategies_by_name = get_test_runner_strategies(num_threads)
+
+ # If the user didn't specify a test runner strategy, determine
+ # the default now based on number of threads and OS type.
+ if not test_runner_name:
+ test_runner_name = default_test_runner_name(num_threads)
+
+ if test_runner_name not in runner_strategies_by_name:
+ raise Exception(
+ "specified testrunner name '{}' unknown. Valid choices: {}".format(
+ test_runner_name,
+ list(runner_strategies_by_name.keys())))
+ test_runner_func = runner_strategies_by_name[test_runner_name]
+
+ # Collect the files on which we'll run the first test run phase.
+ test_files = []
+ find_test_files_in_dir_tree(
+ test_subdir, lambda tdir, tfiles: test_files.append(
+ (test_subdir, tfiles)))
+
+ # Do the first test run phase.
+ summary_results = walk_and_invoke(
+ test_files,
+ dotest_argv,
+ num_threads,
+ test_runner_func)
+
+ (timed_out, passed, failed, unexpected_successes, pass_count,
+ fail_count) = summary_results
+
+ # Check if we have any tests to rerun as phase 2.
+ if results_formatter is not None:
+ tests_for_rerun = results_formatter.tests_for_rerun
+ results_formatter.tests_for_rerun = {}
+
+ if tests_for_rerun is not None and len(tests_for_rerun) > 0:
+ rerun_file_count = len(tests_for_rerun)
+ print("\n{} test files marked for rerun\n".format(
+ rerun_file_count))
+
+ # Check if the number of files exceeds the max cutoff. If so,
+ # we skip the rerun step.
+ if rerun_file_count > configuration.rerun_max_file_threshold:
+ print("Skipping rerun: max rerun file threshold ({}) "
+ "exceeded".format(
+ configuration.rerun_max_file_threshold))
+ else:
+ rerun_tests(test_subdir, tests_for_rerun, dotest_argv)
+
+ # The results formatter - if present - is done now. Tell it to
+ # terminate.
+ if results_formatter is not None:
+ results_formatter.send_terminate_as_needed()
+
+ timed_out = set(timed_out)
+ num_test_files = len(passed) + len(failed)
+ num_test_cases = pass_count + fail_count
+
+ # move core files into session dir
+ cores = find('core.*', test_subdir)
+ for core in cores:
+ dst = core.replace(test_directory, "")[1:]
+ dst = dst.replace(os.path.sep, "-")
+ os.rename(core, os.path.join(session_dir, dst))
+
+ # remove expected timeouts from failures
+ for xtime in expected_timeout:
+ if xtime in timed_out:
+ timed_out.remove(xtime)
+ failed.remove(xtime)
+ result = "ExpectedTimeout"
+ elif xtime in passed:
+ result = "UnexpectedCompletion"
+ else:
+ result = None # failed
+
+ if result:
+ test_name = os.path.splitext(xtime)[0]
+ touch(os.path.join(session_dir, "{}-{}".format(result, test_name)))
+
+ # Only run the old summary logic if we don't have a results formatter
+ # that already prints the summary.
+ print_legacy_summary = results_formatter is None
+ if not print_legacy_summary:
+ # Print summary results. Summarized results at the end always
+ # get printed to stdout, even if --results-file specifies a different
+ # file for, say, xUnit output.
+ results_formatter.print_results(sys.stdout)
+
+ # Figure out exit code by count of test result types.
+ issue_count = 0
+ for issue_status in EventBuilder.TESTRUN_ERROR_STATUS_VALUES:
+ issue_count += results_formatter.counts_by_test_result_status(
+ issue_status)
+
+ # Return with appropriate result code
+ if issue_count > 0:
+ sys.exit(1)
+ else:
+ sys.exit(0)
+ else:
+ # Print the legacy test results summary.
+ print()
+ sys.stdout.write("Ran %d test suites" % num_test_files)
+ if num_test_files > 0:
+ sys.stdout.write(" (%d failed) (%f%%)" % (
+ len(failed), 100.0 * len(failed) / num_test_files))
+ print()
+ sys.stdout.write("Ran %d test cases" % num_test_cases)
+ if num_test_cases > 0:
+ sys.stdout.write(" (%d failed) (%f%%)" % (
+ fail_count, 100.0 * fail_count / num_test_cases))
+ print()
+ exit_code = 0
+
+ if len(failed) > 0:
+ failed.sort()
+ print("Failing Tests (%d)" % len(failed))
+ for f in failed:
+ print("%s: LLDB (suite) :: %s (%s)" % (
+ "TIMEOUT" if f in timed_out else "FAIL", f, system_info
+ ))
+ exit_code = 1
+
+ if len(unexpected_successes) > 0:
+ unexpected_successes.sort()
+ print("\nUnexpected Successes (%d)" % len(unexpected_successes))
+ for u in unexpected_successes:
+ print("UNEXPECTED SUCCESS: LLDB (suite) :: %s (%s)" % (u, system_info))
+
+ sys.exit(exit_code)
+
+if __name__ == '__main__':
+ sys.stderr.write(
+ "error: dosep.py no longer supports being called directly. "
+ "Please call dotest.py directly. The dosep.py-specific arguments "
+ "have been added under the Parallel processing arguments.\n")
+ sys.exit(128)
OpenPOWER on IntegriCloud