diff options
Diffstat (limited to 'packages/Python/lldbsuite/test/dosep.py')
-rw-r--r-- | packages/Python/lldbsuite/test/dosep.py | 1680 |
1 files changed, 1680 insertions, 0 deletions
diff --git a/packages/Python/lldbsuite/test/dosep.py b/packages/Python/lldbsuite/test/dosep.py new file mode 100644 index 0000000..51275d5 --- /dev/null +++ b/packages/Python/lldbsuite/test/dosep.py @@ -0,0 +1,1680 @@ +""" +Run the test suite using a separate process for each test file. + +Each test will run with a time limit of 10 minutes by default. + +Override the default time limit of 10 minutes by setting +the environment variable LLDB_TEST_TIMEOUT. + +E.g., export LLDB_TEST_TIMEOUT=10m + +Override the time limit for individual tests by setting +the environment variable LLDB_[TEST NAME]_TIMEOUT. + +E.g., export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=2m + +Set to "0" to run without time limit. + +E.g., export LLDB_TEST_TIMEOUT=0 +or export LLDB_TESTCONCURRENTEVENTS_TIMEOUT=0 + +To collect core files for timed out tests, +do the following before running dosep.py + +OSX +ulimit -c unlimited +sudo sysctl -w kern.corefile=core.%P + +Linux: +ulimit -c unlimited +echo core.%p | sudo tee /proc/sys/kernel/core_pattern +""" + +from __future__ import print_function +from __future__ import absolute_import + +# system packages and modules +import asyncore +import distutils.version +import fnmatch +import multiprocessing +import multiprocessing.pool +import os +import platform +import re +import signal +import sys +import threading + +from six.moves import queue + +# Our packages and modules +import lldbsuite +import lldbsuite.support.seven as seven + +from . import configuration +from . import dotest_channels +from . import dotest_args +from . import result_formatter + +from .result_formatter import EventBuilder + + +# Todo: Convert this folder layout to be relative-import friendly and +# don't hack up sys.path like this +sys.path.append(os.path.join(os.path.dirname(__file__), "test_runner", "lib")) +import lldb_utils +import process_control + +# Status codes for running command with timeout. +eTimedOut, ePassed, eFailed = 124, 0, 1 + +output_lock = None +test_counter = None +total_tests = None +test_name_len = None +dotest_options = None +RESULTS_FORMATTER = None +RUNNER_PROCESS_ASYNC_MAP = None +RESULTS_LISTENER_CHANNEL = None + +"""Contains an optional function pointer that can return the worker index + for the given thread/process calling it. Returns a 0-based index.""" +GET_WORKER_INDEX = None + + +def setup_global_variables( + lock, counter, total, name_len, options, worker_index_map): + global output_lock, test_counter, total_tests, test_name_len + global dotest_options + output_lock = lock + test_counter = counter + total_tests = total + test_name_len = name_len + dotest_options = options + + if worker_index_map is not None: + # We'll use the output lock for this to avoid sharing another lock. + # This won't be used much. + index_lock = lock + + def get_worker_index_use_pid(): + """Returns a 0-based, process-unique index for the worker.""" + pid = os.getpid() + with index_lock: + if pid not in worker_index_map: + worker_index_map[pid] = len(worker_index_map) + return worker_index_map[pid] + + global GET_WORKER_INDEX + GET_WORKER_INDEX = get_worker_index_use_pid + +def report_test_failure(name, command, output): + global output_lock + with output_lock: + if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()): + print(file=sys.stderr) + print(output, file=sys.stderr) + print("[%s FAILED]" % name, file=sys.stderr) + print("Command invoked: %s" % ' '.join(command), file=sys.stderr) + update_progress(name) + + +def report_test_pass(name, output): + global output_lock + with output_lock: + update_progress(name) + + +def update_progress(test_name=""): + global output_lock, test_counter, total_tests, test_name_len + with output_lock: + counter_len = len(str(total_tests)) + if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()): + sys.stderr.write( + "\r%*d out of %d test suites processed - %-*s" % + (counter_len, test_counter.value, total_tests, + test_name_len.value, test_name)) + if len(test_name) > test_name_len.value: + test_name_len.value = len(test_name) + test_counter.value += 1 + sys.stdout.flush() + sys.stderr.flush() + + +def parse_test_results(output): + passes = 0 + failures = 0 + unexpected_successes = 0 + for result in output: + pass_count = re.search("^RESULT:.*([0-9]+) passes", + result, re.MULTILINE) + fail_count = re.search("^RESULT:.*([0-9]+) failures", + result, re.MULTILINE) + error_count = re.search("^RESULT:.*([0-9]+) errors", + result, re.MULTILINE) + unexpected_success_count = re.search("^RESULT:.*([0-9]+) unexpected successes", + result, re.MULTILINE) + if pass_count is not None: + passes = passes + int(pass_count.group(1)) + if fail_count is not None: + failures = failures + int(fail_count.group(1)) + if unexpected_success_count is not None: + unexpected_successes = unexpected_successes + int(unexpected_success_count.group(1)) + if error_count is not None: + failures = failures + int(error_count.group(1)) + return passes, failures, unexpected_successes + + +class DoTestProcessDriver(process_control.ProcessDriver): + """Drives the dotest.py inferior process and handles bookkeeping.""" + def __init__(self, output_file, output_file_lock, pid_events, file_name, + soft_terminate_timeout): + super(DoTestProcessDriver, self).__init__( + soft_terminate_timeout=soft_terminate_timeout) + self.output_file = output_file + self.output_lock = lldb_utils.OptionalWith(output_file_lock) + self.pid_events = pid_events + self.results = None + self.file_name = file_name + + def write(self, content): + with self.output_lock: + self.output_file.write(content) + + def on_process_started(self): + if self.pid_events: + self.pid_events.put_nowait(('created', self.process.pid)) + + def on_process_exited(self, command, output, was_timeout, exit_status): + if self.pid_events: + # No point in culling out those with no exit_status (i.e. + # those we failed to kill). That would just cause + # downstream code to try to kill it later on a Ctrl-C. At + # this point, a best-effort-to-kill already took place. So + # call it destroyed here. + self.pid_events.put_nowait(('destroyed', self.process.pid)) + + # Override the exit status if it was a timeout. + if was_timeout: + exit_status = eTimedOut + + # If we didn't end up with any output, call it empty for + # stdout/stderr. + if output is None: + output = ('', '') + + # Now parse the output. + passes, failures, unexpected_successes = parse_test_results(output) + if exit_status == 0: + # stdout does not have any useful information from 'dotest.py', + # only stderr does. + report_test_pass(self.file_name, output[1]) + else: + report_test_failure(self.file_name, command, output[1]) + + # Save off the results for the caller. + self.results = ( + self.file_name, + exit_status, + passes, + failures, + unexpected_successes) + + def is_exceptional_exit(self): + """Returns whether the process returned a timeout. + + Not valid to call until after on_process_exited() completes. + + @return True if the exit is an exceptional exit (e.g. signal on + POSIX); False otherwise. + """ + if self.results is None: + raise Exception( + "exit status checked before results are available") + return self.process_helper.is_exceptional_exit( + self.results[1]) + + def exceptional_exit_details(self): + if self.results is None: + raise Exception( + "exit status checked before results are available") + return self.process_helper.exceptional_exit_details(self.results[1]) + + def is_timeout(self): + if self.results is None: + raise Exception( + "exit status checked before results are available") + return self.results[1] == eTimedOut + + +def get_soft_terminate_timeout(): + # Defaults to 10 seconds, but can set + # LLDB_TEST_SOFT_TERMINATE_TIMEOUT to a floating point + # number in seconds. This value indicates how long + # the test runner will wait for the dotest inferior to + # handle a timeout via a soft terminate before it will + # assume that failed and do a hard terminate. + + # TODO plumb through command-line option + return float(os.environ.get('LLDB_TEST_SOFT_TERMINATE_TIMEOUT', 10.0)) + + +def want_core_on_soft_terminate(): + # TODO plumb through command-line option + if platform.system() == 'Linux': + return True + else: + return False + + +def send_events_to_collector(events, command): + """Sends the given events to the collector described in the command line. + + @param events the list of events to send to the test event collector. + @param command the inferior command line which contains the details on + how to connect to the test event collector. + """ + if events is None or len(events) == 0: + # Nothing to do. + return + + # Find the port we need to connect to from the --results-port option. + try: + arg_index = command.index("--results-port") + 1 + except ValueError: + # There is no results port, so no way to communicate back to + # the event collector. This is not a problem if we're not + # using event aggregation. + # TODO flag as error once we always use the event system + print( + "INFO: no event collector, skipping post-inferior test " + "event reporting") + return + + if arg_index >= len(command): + raise Exception( + "expected collector port at index {} in {}".format( + arg_index, command)) + event_port = int(command[arg_index]) + + # Create results formatter connected back to collector via socket. + config = result_formatter.FormatterConfig() + config.port = event_port + formatter_spec = result_formatter.create_results_formatter(config) + if formatter_spec is None or formatter_spec.formatter is None: + raise Exception( + "Failed to create socket-based ResultsFormatter " + "back to test event collector") + + # Send the events: the port-based event just pickles the content + # and sends over to the server side of the socket. + for event in events: + formatter_spec.formatter.handle_event(event) + + # Cleanup + if formatter_spec.cleanup_func is not None: + formatter_spec.cleanup_func() + + +def send_inferior_post_run_events( + command, worker_index, process_driver, test_filename): + """Sends any test events that should be generated after the inferior runs. + + These events would include timeouts and exceptional (i.e. signal-returning) + process completion results. + + @param command the list of command parameters passed to subprocess.Popen(). + @param worker_index the worker index (possibly None) used to run + this process + @param process_driver the ProcessDriver-derived instance that was used + to run the inferior process. + @param test_filename the full path to the Python test file that is being + run. + """ + if process_driver is None: + raise Exception("process_driver must not be None") + if process_driver.results is None: + # Invalid condition - the results should have been set one way or + # another, even in a timeout. + raise Exception("process_driver.results were not set") + + # The code below fills in the post events struct. If there are any post + # events to fire up, we'll try to make a connection to the socket and + # provide the results. + post_events = [] + + # Handle signal/exceptional exits. + if process_driver.is_exceptional_exit(): + (code, desc) = process_driver.exceptional_exit_details() + post_events.append( + EventBuilder.event_for_job_exceptional_exit( + process_driver.pid, + worker_index, + code, + desc, + test_filename, + command)) + + # Handle timeouts. + if process_driver.is_timeout(): + post_events.append(EventBuilder.event_for_job_timeout( + process_driver.pid, + worker_index, + test_filename, + command)) + + if len(post_events) > 0: + send_events_to_collector(post_events, command) + + +def call_with_timeout( + command, timeout, name, inferior_pid_events, test_filename): + # Add our worker index (if we have one) to all test events + # from this inferior. + worker_index = None + if GET_WORKER_INDEX is not None: + try: + worker_index = GET_WORKER_INDEX() + command.extend([ + "--event-add-entries", + "worker_index={}:int".format(worker_index)]) + except: # pylint: disable=bare-except + # Ctrl-C does bad things to multiprocessing.Manager.dict() + # lookup. Just swallow it. + pass + + # Create the inferior dotest.py ProcessDriver. + soft_terminate_timeout = get_soft_terminate_timeout() + want_core = want_core_on_soft_terminate() + + process_driver = DoTestProcessDriver( + sys.stdout, + output_lock, + inferior_pid_events, + name, + soft_terminate_timeout) + + # Run it with a timeout. + process_driver.run_command_with_timeout(command, timeout, want_core) + + # Return the results. + if not process_driver.results: + # This is truly exceptional. Even a failing or timed out + # binary should have called the results-generation code. + raise Exception("no test results were generated whatsoever") + + # Handle cases where the test inferior cannot adequately provide + # meaningful results to the test event system. + send_inferior_post_run_events( + command, + worker_index, + process_driver, + test_filename) + + return process_driver.results + + +def process_dir(root, files, dotest_argv, inferior_pid_events): + """Examine a directory for tests, and invoke any found within it.""" + results = [] + for (base_name, full_test_path) in files: + import __main__ as main + script_file = main.__file__ + command = ([sys.executable, script_file] + + dotest_argv + + ["--inferior", "-p", base_name, root]) + + timeout_name = os.path.basename(os.path.splitext(base_name)[0]).upper() + + timeout = (os.getenv("LLDB_%s_TIMEOUT" % timeout_name) or + getDefaultTimeout(dotest_options.lldb_platform_name)) + + results.append(call_with_timeout( + command, timeout, base_name, inferior_pid_events, full_test_path)) + + # result = (name, status, passes, failures, unexpected_successes) + timed_out = [name for name, status, _, _, _ in results + if status == eTimedOut] + passed = [name for name, status, _, _, _ in results + if status == ePassed] + failed = [name for name, status, _, _, _ in results + if status != ePassed] + unexpected_passes = [ + name for name, _, _, _, unexpected_successes in results + if unexpected_successes > 0] + + pass_count = sum([result[2] for result in results]) + fail_count = sum([result[3] for result in results]) + + return ( + timed_out, passed, failed, unexpected_passes, pass_count, fail_count) + +in_q = None +out_q = None + + +def process_dir_worker_multiprocessing( + a_output_lock, a_test_counter, a_total_tests, a_test_name_len, + a_dotest_options, job_queue, result_queue, inferior_pid_events, + worker_index_map): + """Worker thread main loop when in multiprocessing mode. + Takes one directory specification at a time and works on it.""" + + # Shut off interrupt handling in the child process. + signal.signal(signal.SIGINT, signal.SIG_IGN) + if hasattr(signal, 'SIGHUP'): + signal.signal(signal.SIGHUP, signal.SIG_IGN) + + # Setup the global state for the worker process. + setup_global_variables( + a_output_lock, a_test_counter, a_total_tests, a_test_name_len, + a_dotest_options, worker_index_map) + + # Keep grabbing entries from the queue until done. + while not job_queue.empty(): + try: + job = job_queue.get(block=False) + result = process_dir(job[0], job[1], job[2], + inferior_pid_events) + result_queue.put(result) + except queue.Empty: + # Fine, we're done. + pass + + +def process_dir_worker_multiprocessing_pool(args): + return process_dir(*args) + + +def process_dir_worker_threading(job_queue, result_queue, inferior_pid_events): + """Worker thread main loop when in threading mode. + + This one supports the hand-rolled pooling support. + + Takes one directory specification at a time and works on it.""" + + # Keep grabbing entries from the queue until done. + while not job_queue.empty(): + try: + job = job_queue.get(block=False) + result = process_dir(job[0], job[1], job[2], + inferior_pid_events) + result_queue.put(result) + except queue.Empty: + # Fine, we're done. + pass + + +def process_dir_worker_threading_pool(args): + return process_dir(*args) + + +def process_dir_mapper_inprocess(args): + """Map adapter for running the subprocess-based, non-threaded test runner. + + @param args the process work item tuple + @return the test result tuple + """ + return process_dir(*args) + + +def collect_active_pids_from_pid_events(event_queue): + """ + Returns the set of what should be active inferior pids based on + the event stream. + + @param event_queue a multiprocessing.Queue containing events of the + form: + ('created', pid) + ('destroyed', pid) + + @return set of inferior dotest.py pids activated but never completed. + """ + active_pid_set = set() + while not event_queue.empty(): + pid_event = event_queue.get_nowait() + if pid_event[0] == 'created': + active_pid_set.add(pid_event[1]) + elif pid_event[0] == 'destroyed': + active_pid_set.remove(pid_event[1]) + return active_pid_set + + +def kill_all_worker_processes(workers, inferior_pid_events): + """ + Kills all specified worker processes and their process tree. + + @param workers a list of multiprocess.Process worker objects. + @param inferior_pid_events a multiprocess.Queue that contains + all inferior create and destroy events. Used to construct + the list of child pids still outstanding that need to be killed. + """ + for worker in workers: + worker.terminate() + worker.join() + + # Add all the child test pids created. + active_pid_set = collect_active_pids_from_pid_events( + inferior_pid_events) + for inferior_pid in active_pid_set: + print("killing inferior pid {}".format(inferior_pid)) + os.kill(inferior_pid, signal.SIGKILL) + + +def kill_all_worker_threads(workers, inferior_pid_events): + """ + Kills all specified worker threads and their process tree. + + @param workers a list of multiprocess.Process worker objects. + @param inferior_pid_events a multiprocess.Queue that contains + all inferior create and destroy events. Used to construct + the list of child pids still outstanding that need to be killed. + """ + + # Add all the child test pids created. + active_pid_set = collect_active_pids_from_pid_events( + inferior_pid_events) + for inferior_pid in active_pid_set: + print("killing inferior pid {}".format(inferior_pid)) + os.kill(inferior_pid, signal.SIGKILL) + + # We don't have a way to nuke the threads. However, since we killed + # all the inferiors, and we drained the job queue, this will be + # good enough. Wait cleanly for each worker thread to wrap up. + for worker in workers: + worker.join() + + +def find_test_files_in_dir_tree(dir_root, found_func): + """Calls found_func for all the test files in the given dir hierarchy. + + @param dir_root the path to the directory to start scanning + for test files. All files in this directory and all its children + directory trees will be searched. + + @param found_func a callable object that will be passed + the parent directory (relative to dir_root) and the list of + test files from within that directory. + """ + for root, _, files in os.walk(dir_root, topdown=False): + def is_test_filename(test_dir, base_filename): + """Returns True if the given filename matches the test name format. + + @param test_dir the directory to check. Should be absolute or + relative to current working directory. + + @param base_filename the base name of the filename to check for a + dherence to the python test case filename format. + + @return True if name matches the python test case filename format. + """ + # Not interested in symbolically linked files. + if os.path.islink(os.path.join(test_dir, base_filename)): + return False + # Only interested in test files with the "Test*.py" naming pattern. + return (base_filename.startswith("Test") and + base_filename.endswith(".py")) + + tests = [ + (filename, os.path.join(root, filename)) + for filename in files + if is_test_filename(root, filename)] + if tests: + found_func(root, tests) + + +def initialize_global_vars_common(num_threads, test_work_items): + global total_tests, test_counter, test_name_len + + total_tests = sum([len(item[1]) for item in test_work_items]) + test_counter = multiprocessing.Value('i', 0) + test_name_len = multiprocessing.Value('i', 0) + if not (RESULTS_FORMATTER and RESULTS_FORMATTER.is_using_terminal()): + print("Testing: %d test suites, %d thread%s" % ( + total_tests, num_threads, (num_threads > 1) * "s"), file=sys.stderr) + update_progress() + + +def initialize_global_vars_multiprocessing(num_threads, test_work_items): + # Initialize the global state we'll use to communicate with the + # rest of the flat module. + global output_lock + output_lock = multiprocessing.RLock() + + initialize_global_vars_common(num_threads, test_work_items) + + +def initialize_global_vars_threading(num_threads, test_work_items): + """Initializes global variables used in threading mode. + @param num_threads specifies the number of workers used. + @param test_work_items specifies all the work items + that will be processed. + """ + # Initialize the global state we'll use to communicate with the + # rest of the flat module. + global output_lock + output_lock = threading.RLock() + + index_lock = threading.RLock() + index_map = {} + + def get_worker_index_threading(): + """Returns a 0-based, thread-unique index for the worker thread.""" + thread_id = threading.current_thread().ident + with index_lock: + if thread_id not in index_map: + index_map[thread_id] = len(index_map) + return index_map[thread_id] + + + global GET_WORKER_INDEX + GET_WORKER_INDEX = get_worker_index_threading + + initialize_global_vars_common(num_threads, test_work_items) + + +def ctrl_c_loop(main_op_func, done_func, ctrl_c_handler): + """Provides a main loop that is Ctrl-C protected. + + The main loop calls the main_op_func() repeatedly until done_func() + returns true. The ctrl_c_handler() method is called with a single + int parameter that contains the number of times the ctrl_c has been + hit (starting with 1). The ctrl_c_handler() should mutate whatever + it needs to have the done_func() return True as soon as it is desired + to exit the loop. + """ + done = False + ctrl_c_count = 0 + + while not done: + try: + # See if we're done. Start with done check since it is + # the first thing executed after a Ctrl-C handler in the + # following loop. + done = done_func() + if not done: + # Run the main op once. + main_op_func() + + except KeyboardInterrupt: + ctrl_c_count += 1 + ctrl_c_handler(ctrl_c_count) + + +def pump_workers_and_asyncore_map(workers, asyncore_map): + """Prunes out completed workers and maintains the asyncore loop. + + The asyncore loop contains the optional socket listener + and handlers. When all workers are complete, this method + takes care of stopping the listener. It also runs the + asyncore loop for the given async map for 10 iterations. + + @param workers the list of worker Thread/Process instances. + + @param asyncore_map the asyncore threading-aware map that + indicates which channels are in use and still alive. + """ + + # Check on all the workers, removing them from the workers + # list as they complete. + dead_workers = [] + for worker in workers: + # This non-blocking join call is what allows us + # to still receive keyboard interrupts. + worker.join(0.01) + if not worker.is_alive(): + dead_workers.append(worker) + # Clear out the completed workers + for dead_worker in dead_workers: + workers.remove(dead_worker) + + # If there are no more workers and there is a listener, + # close the listener. + global RESULTS_LISTENER_CHANNEL + if len(workers) == 0 and RESULTS_LISTENER_CHANNEL is not None: + RESULTS_LISTENER_CHANNEL.close() + RESULTS_LISTENER_CHANNEL = None + + # Pump the asyncore map if it isn't empty. + if len(asyncore_map) > 0: + asyncore.loop(0.1, False, asyncore_map, 10) + + +def handle_ctrl_c(ctrl_c_count, job_queue, workers, inferior_pid_events, + stop_all_inferiors_func): + """Performs the appropriate ctrl-c action for non-pool parallel test runners + + @param ctrl_c_count starting with 1, indicates the number of times ctrl-c + has been intercepted. The value is 1 on the first intercept, 2 on the + second, etc. + + @param job_queue a Queue object that contains the work still outstanding + (i.e. hasn't been assigned to a worker yet). + + @param workers list of Thread or Process workers. + + @param inferior_pid_events specifies a Queue of inferior process + construction and destruction events. Used to build the list of inferior + processes that should be killed if we get that far. + + @param stop_all_inferiors_func a callable object that takes the + workers and inferior_pid_events parameters (in that order) if a hard + stop is to be used on the workers. + """ + + # Print out which Ctrl-C we're handling. + key_name = [ + "first", + "second", + "third", + "many"] + + if ctrl_c_count < len(key_name): + name_index = ctrl_c_count - 1 + else: + name_index = len(key_name) - 1 + message = "\nHandling {} KeyboardInterrupt".format(key_name[name_index]) + with output_lock: + print(message) + + if ctrl_c_count == 1: + # Remove all outstanding items from the work queue so we stop + # doing any more new work. + while not job_queue.empty(): + try: + # Just drain it to stop more work from being started. + job_queue.get_nowait() + except queue.Empty: + pass + with output_lock: + print("Stopped more work from being started.") + elif ctrl_c_count == 2: + # Try to stop all inferiors, even the ones currently doing work. + stop_all_inferiors_func(workers, inferior_pid_events) + else: + with output_lock: + print("All teardown activities kicked off, should finish soon.") + + +def workers_and_async_done(workers, async_map): + """Returns True if the workers list and asyncore channels are all done. + + @param workers list of workers (threads/processes). These must adhere + to the threading Thread or multiprocessing.Process interface. + + @param async_map the threading-aware asyncore channel map to check + for live channels. + + @return False if the workers list exists and has any entries in it, or + if the async_map exists and has any entries left in it; otherwise, True. + """ + if workers is not None and len(workers) > 0: + # We're not done if we still have workers left. + return False + if async_map is not None and len(async_map) > 0: + return False + # We're done. + return True + + +def multiprocessing_test_runner(num_threads, test_work_items): + """Provides hand-wrapped pooling test runner adapter with Ctrl-C support. + + This concurrent test runner is based on the multiprocessing + library, and rolls its own worker pooling strategy so it + can handle Ctrl-C properly. + + This test runner is known to have an issue running on + Windows platforms. + + @param num_threads the number of worker processes to use. + + @param test_work_items the iterable of test work item tuples + to run. + """ + + # Initialize our global state. + initialize_global_vars_multiprocessing(num_threads, test_work_items) + + # Create jobs. + job_queue = multiprocessing.Queue(len(test_work_items)) + for test_work_item in test_work_items: + job_queue.put(test_work_item) + + result_queue = multiprocessing.Queue(len(test_work_items)) + + # Create queues for started child pids. Terminating + # the multiprocess processes does not terminate the + # child processes they spawn. We can remove this tracking + # if/when we move to having the multiprocess process directly + # perform the test logic. The Queue size needs to be able to + # hold 2 * (num inferior dotest.py processes started) entries. + inferior_pid_events = multiprocessing.Queue(4096) + + # Worker dictionary allows each worker to figure out its worker index. + manager = multiprocessing.Manager() + worker_index_map = manager.dict() + + # Create workers. We don't use multiprocessing.Pool due to + # challenges with handling ^C keyboard interrupts. + workers = [] + for _ in range(num_threads): + worker = multiprocessing.Process( + target=process_dir_worker_multiprocessing, + args=(output_lock, + test_counter, + total_tests, + test_name_len, + dotest_options, + job_queue, + result_queue, + inferior_pid_events, + worker_index_map)) + worker.start() + workers.append(worker) + + # Main loop: wait for all workers to finish and wait for + # the socket handlers to wrap up. + ctrl_c_loop( + # Main operation of loop + lambda: pump_workers_and_asyncore_map( + workers, RUNNER_PROCESS_ASYNC_MAP), + + # Return True when we're done with the main loop. + lambda: workers_and_async_done(workers, RUNNER_PROCESS_ASYNC_MAP), + + # Indicate what we do when we receive one or more Ctrl-Cs. + lambda ctrl_c_count: handle_ctrl_c( + ctrl_c_count, job_queue, workers, inferior_pid_events, + kill_all_worker_processes)) + + # Reap the test results. + test_results = [] + while not result_queue.empty(): + test_results.append(result_queue.get(block=False)) + return test_results + + +def map_async_run_loop(future, channel_map, listener_channel): + """Blocks until the Pool.map_async completes and the channel completes. + + @param future an AsyncResult instance from a Pool.map_async() call. + + @param channel_map the asyncore dispatch channel map that should be pumped. + Optional: may be None. + + @param listener_channel the channel representing a listener that should be + closed once the map_async results are available. + + @return the results from the async_result instance. + """ + map_results = None + + done = False + while not done: + # Check if we need to reap the map results. + if map_results is None: + if future.ready(): + # Get the results. + map_results = future.get() + + # Close the runner process listener channel if we have + # one: no more connections will be incoming. + if listener_channel is not None: + listener_channel.close() + + # Pump the asyncore loop if we have a listener socket. + if channel_map is not None: + asyncore.loop(0.01, False, channel_map, 10) + + # Figure out if we're done running. + done = map_results is not None + if channel_map is not None: + # We have a runner process async map. Check if it + # is complete. + if len(channel_map) > 0: + # We still have an asyncore channel running. Not done yet. + done = False + + return map_results + + +def multiprocessing_test_runner_pool(num_threads, test_work_items): + # Initialize our global state. + initialize_global_vars_multiprocessing(num_threads, test_work_items) + + manager = multiprocessing.Manager() + worker_index_map = manager.dict() + + pool = multiprocessing.Pool( + num_threads, + initializer=setup_global_variables, + initargs=(output_lock, test_counter, total_tests, test_name_len, + dotest_options, worker_index_map)) + + # Start the map operation (async mode). + map_future = pool.map_async( + process_dir_worker_multiprocessing_pool, test_work_items) + return map_async_run_loop( + map_future, RUNNER_PROCESS_ASYNC_MAP, RESULTS_LISTENER_CHANNEL) + + +def threading_test_runner(num_threads, test_work_items): + """Provides hand-wrapped pooling threading-based test runner adapter + with Ctrl-C support. + + This concurrent test runner is based on the threading + library, and rolls its own worker pooling strategy so it + can handle Ctrl-C properly. + + @param num_threads the number of worker processes to use. + + @param test_work_items the iterable of test work item tuples + to run. + """ + + # Initialize our global state. + initialize_global_vars_threading(num_threads, test_work_items) + + # Create jobs. + job_queue = queue.Queue() + for test_work_item in test_work_items: + job_queue.put(test_work_item) + + result_queue = queue.Queue() + + # Create queues for started child pids. Terminating + # the threading threads does not terminate the + # child processes they spawn. + inferior_pid_events = queue.Queue() + + # Create workers. We don't use multiprocessing.pool.ThreadedPool + # due to challenges with handling ^C keyboard interrupts. + workers = [] + for _ in range(num_threads): + worker = threading.Thread( + target=process_dir_worker_threading, + args=(job_queue, + result_queue, + inferior_pid_events)) + worker.start() + workers.append(worker) + + # Main loop: wait for all workers to finish and wait for + # the socket handlers to wrap up. + ctrl_c_loop( + # Main operation of loop + lambda: pump_workers_and_asyncore_map( + workers, RUNNER_PROCESS_ASYNC_MAP), + + # Return True when we're done with the main loop. + lambda: workers_and_async_done(workers, RUNNER_PROCESS_ASYNC_MAP), + + # Indicate what we do when we receive one or more Ctrl-Cs. + lambda ctrl_c_count: handle_ctrl_c( + ctrl_c_count, job_queue, workers, inferior_pid_events, + kill_all_worker_threads)) + + # Reap the test results. + test_results = [] + while not result_queue.empty(): + test_results.append(result_queue.get(block=False)) + return test_results + + +def threading_test_runner_pool(num_threads, test_work_items): + # Initialize our global state. + initialize_global_vars_threading(num_threads, test_work_items) + + pool = multiprocessing.pool.ThreadPool(num_threads) + map_future = pool.map_async( + process_dir_worker_threading_pool, test_work_items) + + return map_async_run_loop( + map_future, RUNNER_PROCESS_ASYNC_MAP, RESULTS_LISTENER_CHANNEL) + + +def asyncore_run_loop(channel_map): + try: + asyncore.loop(None, False, channel_map) + except: + # Swallow it, we're seeing: + # error: (9, 'Bad file descriptor') + # when the listener channel is closed. Shouldn't be the case. + pass + + +def inprocess_exec_test_runner(test_work_items): + # Initialize our global state. + initialize_global_vars_multiprocessing(1, test_work_items) + + # We're always worker index 0 + global GET_WORKER_INDEX + GET_WORKER_INDEX = lambda: 0 + + # Run the listener and related channel maps in a separate thread. + # global RUNNER_PROCESS_ASYNC_MAP + global RESULTS_LISTENER_CHANNEL + if RESULTS_LISTENER_CHANNEL is not None: + socket_thread = threading.Thread( + target=lambda: asyncore_run_loop(RUNNER_PROCESS_ASYNC_MAP)) + socket_thread.start() + + # Do the work. + test_results = list(map(process_dir_mapper_inprocess, test_work_items)) + + # If we have a listener channel, shut it down here. + if RESULTS_LISTENER_CHANNEL is not None: + # Close down the channel. + RESULTS_LISTENER_CHANNEL.close() + RESULTS_LISTENER_CHANNEL = None + + # Wait for the listener and handlers to complete. + socket_thread.join() + + return test_results + +def walk_and_invoke(test_files, dotest_argv, num_workers, test_runner_func): + """Invokes the test runner on each test file specified by test_files. + + @param test_files a list of (test_subdir, list_of_test_files_in_dir) + @param num_workers the number of worker queues working on these test files + @param test_runner_func the test runner configured to run the tests + + @return a tuple of results from the running of the specified tests, + of the form (timed_out, passed, failed, unexpected_successes, pass_count, + fail_count) + """ + # The async_map is important to keep all thread-related asyncore + # channels distinct when we call asyncore.loop() later on. + global RESULTS_LISTENER_CHANNEL, RUNNER_PROCESS_ASYNC_MAP + RUNNER_PROCESS_ASYNC_MAP = {} + + # If we're outputting side-channel test results, create the socket + # listener channel and tell the inferior to send results to the + # port on which we'll be listening. + if RESULTS_FORMATTER is not None: + forwarding_func = RESULTS_FORMATTER.handle_event + RESULTS_LISTENER_CHANNEL = ( + dotest_channels.UnpicklingForwardingListenerChannel( + RUNNER_PROCESS_ASYNC_MAP, "localhost", 0, + 2 * num_workers, forwarding_func)) + # Set the results port command line arg. Might have been + # inserted previous, so first try to replace. + listener_port = str(RESULTS_LISTENER_CHANNEL.address[1]) + try: + port_value_index = dotest_argv.index("--results-port") + 1 + dotest_argv[port_value_index] = listener_port + except ValueError: + # --results-port doesn't exist (yet), add it + dotest_argv.append("--results-port") + dotest_argv.append(listener_port) + + # Build the test work items out of the (dir, file_list) entries passed in. + test_work_items = [] + for entry in test_files: + test_work_items.append((entry[0], entry[1], dotest_argv, None)) + + # Convert test work items into test results using whatever + # was provided as the test run function. + test_results = test_runner_func(test_work_items) + + # Summarize the results and return to caller. + timed_out = sum([result[0] for result in test_results], []) + passed = sum([result[1] for result in test_results], []) + failed = sum([result[2] for result in test_results], []) + unexpected_successes = sum([result[3] for result in test_results], []) + pass_count = sum([result[4] for result in test_results]) + fail_count = sum([result[5] for result in test_results]) + + return (timed_out, passed, failed, unexpected_successes, pass_count, + fail_count) + + +def getExpectedTimeouts(platform_name): + # returns a set of test filenames that might timeout + # are we running against a remote target? + host = sys.platform + if platform_name is None: + target = sys.platform + else: + m = re.search(r'remote-(\w+)', platform_name) + target = m.group(1) + + expected_timeout = set() + + if target.startswith("android"): + expected_timeout |= { + "TestExitDuringStep.py", + "TestHelloWorld.py", + } + if host.startswith("win32"): + expected_timeout |= { + "TestEvents.py", + "TestThreadStates.py", + } + elif target.startswith("freebsd"): + expected_timeout |= { + "TestBreakpointConditions.py", + "TestChangeProcessGroup.py", + "TestValueObjectRecursion.py", + "TestWatchpointConditionAPI.py", + } + elif target.startswith("darwin"): + expected_timeout |= { + # times out on MBP Retina, Mid 2012 + "TestThreadSpecificBreakpoint.py", + "TestExitDuringStep.py", + "TestIntegerTypesExpr.py", + } + return expected_timeout + + +def getDefaultTimeout(platform_name): + if os.getenv("LLDB_TEST_TIMEOUT"): + return os.getenv("LLDB_TEST_TIMEOUT") + + if platform_name is None: + platform_name = sys.platform + + if platform_name.startswith("remote-"): + return "10m" + elif platform_name == 'darwin': + # We are consistently needing more time on a few tests. + return "6m" + else: + return "4m" + + +def touch(fname, times=None): + if os.path.exists(fname): + os.utime(fname, times) + + +def find(pattern, path): + result = [] + for root, dirs, files in os.walk(path): + for name in files: + if fnmatch.fnmatch(name, pattern): + result.append(os.path.join(root, name)) + return result + + +def get_test_runner_strategies(num_threads): + """Returns the test runner strategies by name in a dictionary. + + @param num_threads specifies the number of threads/processes + that will be used for concurrent test runners. + + @return dictionary with key as test runner strategy name and + value set to a callable object that takes the test work item + and returns a test result tuple. + """ + return { + # multiprocessing supports ctrl-c and does not use + # multiprocessing.Pool. + "multiprocessing": + (lambda work_items: multiprocessing_test_runner( + num_threads, work_items)), + + # multiprocessing-pool uses multiprocessing.Pool but + # does not support Ctrl-C. + "multiprocessing-pool": + (lambda work_items: multiprocessing_test_runner_pool( + num_threads, work_items)), + + # threading uses a hand-rolled worker pool much + # like multiprocessing, but instead uses in-process + # worker threads. This one supports Ctrl-C. + "threading": + (lambda work_items: threading_test_runner(num_threads, work_items)), + + # threading-pool uses threading for the workers (in-process) + # and uses the multiprocessing.pool thread-enabled pool. + # This does not properly support Ctrl-C. + "threading-pool": + (lambda work_items: threading_test_runner_pool( + num_threads, work_items)), + + # serial uses the subprocess-based, single process + # test runner. This provides process isolation but + # no concurrent test execution. + "serial": + inprocess_exec_test_runner + } + + +def _remove_option( + args, long_option_name, short_option_name, takes_arg): + """Removes option and related option arguments from args array. + + This method removes all short/long options that match the given + arguments. + + @param args the array of command line arguments (in/out) + + @param long_option_name the full command line representation of the + long-form option that will be removed (including '--'). + + @param short_option_name the short version of the command line option + that will be removed (including '-'). + + @param takes_arg True if the option takes an argument. + + """ + if long_option_name is not None: + regex_string = "^" + long_option_name + "=" + long_regex = re.compile(regex_string) + if short_option_name is not None: + # Short options we only match the -X and assume + # any arg is one command line argument jammed together. + # i.e. -O--abc=1 is a single argument in the args list. + # We don't handle -O --abc=1, as argparse doesn't handle + # it, either. + regex_string = "^" + short_option_name + short_regex = re.compile(regex_string) + + def remove_long_internal(): + """Removes one matching long option from args. + @returns True if one was found and removed; False otherwise. + """ + try: + index = args.index(long_option_name) + # Handle the exact match case. + if takes_arg: + removal_count = 2 + else: + removal_count = 1 + del args[index:index+removal_count] + return True + except ValueError: + # Thanks to argparse not handling options with known arguments + # like other options parsing libraries (see + # https://bugs.python.org/issue9334), we need to support the + # --results-formatter-options={second-level-arguments} (note + # the equal sign to fool the first-level arguments parser into + # not treating the second-level arguments as first-level + # options). We're certainly at risk of getting this wrong + # since now we're forced into the business of trying to figure + # out what is an argument (although I think this + # implementation will suffice). + for index in range(len(args)): + match = long_regex.search(args[index]) + if match: + del args[index] + return True + return False + + def remove_short_internal(): + """Removes one matching short option from args. + @returns True if one was found and removed; False otherwise. + """ + for index in range(len(args)): + match = short_regex.search(args[index]) + if match: + del args[index] + return True + return False + + removal_count = 0 + while long_option_name is not None and remove_long_internal(): + removal_count += 1 + while short_option_name is not None and remove_short_internal(): + removal_count += 1 + if removal_count == 0: + raise Exception( + "failed to find at least one of '{}', '{}' in options".format( + long_option_name, short_option_name)) + + +def adjust_inferior_options(dotest_argv): + """Adjusts the commandline args array for inferiors. + + This method adjusts the inferior dotest commandline options based + on the parallel test runner's options. Some of the inferior options + will need to change to properly handle aggregation functionality. + """ + global dotest_options + + # If we don't have a session directory, create one. + if not dotest_options.s: + # no session log directory, we need to add this to prevent + # every dotest invocation from creating its own directory + import datetime + # The windows platforms don't like ':' in the pathname. + timestamp_started = datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S") + dotest_argv.append('-s') + dotest_argv.append(timestamp_started) + dotest_options.s = timestamp_started + + # Adjust inferior results formatter options - if the parallel + # test runner is collecting into the user-specified test results, + # we'll have inferiors spawn with the --results-port option and + # strip the original test runner options. + if dotest_options.results_file is not None: + _remove_option(dotest_argv, "--results-file", None, True) + if dotest_options.results_port is not None: + _remove_option(dotest_argv, "--results-port", None, True) + if dotest_options.results_formatter is not None: + _remove_option(dotest_argv, "--results-formatter", None, True) + if dotest_options.results_formatter_options is not None: + _remove_option(dotest_argv, "--results-formatter-option", "-O", + True) + + # Remove the --curses shortcut if specified. + if dotest_options.curses: + _remove_option(dotest_argv, "--curses", None, False) + + # Remove test runner name if present. + if dotest_options.test_runner_name is not None: + _remove_option(dotest_argv, "--test-runner-name", None, True) + + +def is_darwin_version_lower_than(target_version): + """Checks that os is Darwin and version is lower than target_version. + + @param target_version the StrictVersion indicating the version + we're checking against. + + @return True if the OS is Darwin (OS X) and the version number of + the OS is less than target_version; False in all other cases. + """ + if platform.system() != 'Darwin': + # Can't be Darwin lower than a certain version. + return False + + system_version = distutils.version.StrictVersion(platform.mac_ver()[0]) + return seven.cmp_(system_version, target_version) < 0 + + +def default_test_runner_name(num_threads): + """Returns the default test runner name for the configuration. + + @param num_threads the number of threads/workers this test runner is + supposed to use. + + @return the test runner name that should be used by default when + no test runner was explicitly called out on the command line. + """ + if num_threads == 1: + # Use the serial runner. + test_runner_name = "serial" + elif os.name == "nt": + # On Windows, Python uses CRT with a low limit on the number of open + # files. If you have a lot of cores, the threading-pool runner will + # often fail because it exceeds that limit. It's not clear what the + # right balance is, so until we can investigate it more deeply, + # just use the one that works + test_runner_name = "multiprocessing-pool" + elif is_darwin_version_lower_than( + distutils.version.StrictVersion("10.10.0")): + # OS X versions before 10.10 appear to have an issue using + # the threading test runner. Fall back to multiprocessing. + # Supports Ctrl-C. + test_runner_name = "multiprocessing" + else: + # For everyone else, use the ctrl-c-enabled threading support. + # Should use fewer system resources than the multprocessing + # variant. + test_runner_name = "threading" + return test_runner_name + + +def rerun_tests(test_subdir, tests_for_rerun, dotest_argv): + # Build the list of test files to rerun. Some future time we'll + # enable re-run by test method so we can constrain the rerun set + # to just the method(s) that were in issued within a file. + + # Sort rerun files into subdirectories. + print("\nRerunning the following files:") + rerun_files_by_subdir = {} + for test_filename in tests_for_rerun.keys(): + # Print the file we'll be rerunning + test_relative_path = os.path.relpath( + test_filename, lldbsuite.lldb_test_root) + print(" {}".format(test_relative_path)) + + # Store test filenames by subdir. + test_dir = os.path.dirname(test_filename) + test_basename = os.path.basename(test_filename) + if test_dir in rerun_files_by_subdir: + rerun_files_by_subdir[test_dir].append( + (test_basename, test_filename)) + else: + rerun_files_by_subdir[test_dir] = [(test_basename, test_filename)] + + # Break rerun work up by subdirectory. We do this since + # we have an invariant that states only one test file can + # be run at a time in any given subdirectory (related to + # rules around built inferior test program lifecycle). + rerun_work = [] + for files_by_subdir in rerun_files_by_subdir.values(): + rerun_work.append((test_subdir, files_by_subdir)) + + # Run the work with the serial runner. + # Do not update legacy counts, I am getting rid of + # them so no point adding complicated merge logic here. + rerun_thread_count = 1 + # Force the parallel test runner to choose a multi-worker strategy. + rerun_runner_name = default_test_runner_name(rerun_thread_count + 1) + print("rerun will use the '{}' test runner strategy".format( + rerun_runner_name)) + + runner_strategies_by_name = get_test_runner_strategies(rerun_thread_count) + rerun_runner_func = runner_strategies_by_name[ + rerun_runner_name] + if rerun_runner_func is None: + raise Exception( + "failed to find rerun test runner " + "function named '{}'".format(rerun_runner_name)) + + walk_and_invoke( + rerun_work, + dotest_argv, + rerun_thread_count, + rerun_runner_func) + print("\nTest rerun complete\n") + + +def main(num_threads, test_subdir, test_runner_name, results_formatter): + """Run dotest.py in inferior mode in parallel. + + @param num_threads the parsed value of the num-threads command line + argument. + + @param test_subdir optionally specifies a subdir to limit testing + within. May be None if the entire test tree is to be used. This subdir + is assumed to be relative to the lldb/test root of the test hierarchy. + + @param test_runner_name if specified, contains the test runner + name which selects the strategy used to run the isolated and + optionally concurrent test runner. Specify None to allow the + system to choose the most appropriate test runner given desired + thread count and OS type. + + @param results_formatter if specified, provides the TestResultsFormatter + instance that will format and output test result data from the + side-channel test results. When specified, inferior dotest calls + will send test results side-channel data over a socket to the parallel + test runner, which will forward them on to results_formatter. + """ + + # Do not shut down on sighup. + if hasattr(signal, 'SIGHUP'): + signal.signal(signal.SIGHUP, signal.SIG_IGN) + + dotest_argv = sys.argv[1:] + + global RESULTS_FORMATTER + RESULTS_FORMATTER = results_formatter + + # We can't use sys.path[0] to determine the script directory + # because it doesn't work under a debugger + parser = dotest_args.create_parser() + global dotest_options + dotest_options = dotest_args.parse_args(parser, dotest_argv) + + adjust_inferior_options(dotest_argv) + + session_dir = os.path.join(os.getcwd(), dotest_options.s) + + # The root directory was specified on the command line + test_directory = os.path.dirname(os.path.realpath(__file__)) + if test_subdir and len(test_subdir) > 0: + test_subdir = os.path.join(test_directory, test_subdir) + else: + test_subdir = test_directory + + # clean core files in test tree from previous runs (Linux) + cores = find('core.*', test_subdir) + for core in cores: + os.unlink(core) + + system_info = " ".join(platform.uname()) + + # Figure out which test files should be enabled for expected + # timeout + expected_timeout = getExpectedTimeouts(dotest_options.lldb_platform_name) + if results_formatter is not None: + results_formatter.set_expected_timeouts_by_basename(expected_timeout) + + # Figure out which testrunner strategy we'll use. + runner_strategies_by_name = get_test_runner_strategies(num_threads) + + # If the user didn't specify a test runner strategy, determine + # the default now based on number of threads and OS type. + if not test_runner_name: + test_runner_name = default_test_runner_name(num_threads) + + if test_runner_name not in runner_strategies_by_name: + raise Exception( + "specified testrunner name '{}' unknown. Valid choices: {}".format( + test_runner_name, + list(runner_strategies_by_name.keys()))) + test_runner_func = runner_strategies_by_name[test_runner_name] + + # Collect the files on which we'll run the first test run phase. + test_files = [] + find_test_files_in_dir_tree( + test_subdir, lambda tdir, tfiles: test_files.append( + (test_subdir, tfiles))) + + # Do the first test run phase. + summary_results = walk_and_invoke( + test_files, + dotest_argv, + num_threads, + test_runner_func) + + (timed_out, passed, failed, unexpected_successes, pass_count, + fail_count) = summary_results + + # Check if we have any tests to rerun as phase 2. + if results_formatter is not None: + tests_for_rerun = results_formatter.tests_for_rerun + results_formatter.tests_for_rerun = {} + + if tests_for_rerun is not None and len(tests_for_rerun) > 0: + rerun_file_count = len(tests_for_rerun) + print("\n{} test files marked for rerun\n".format( + rerun_file_count)) + + # Check if the number of files exceeds the max cutoff. If so, + # we skip the rerun step. + if rerun_file_count > configuration.rerun_max_file_threshold: + print("Skipping rerun: max rerun file threshold ({}) " + "exceeded".format( + configuration.rerun_max_file_threshold)) + else: + rerun_tests(test_subdir, tests_for_rerun, dotest_argv) + + # The results formatter - if present - is done now. Tell it to + # terminate. + if results_formatter is not None: + results_formatter.send_terminate_as_needed() + + timed_out = set(timed_out) + num_test_files = len(passed) + len(failed) + num_test_cases = pass_count + fail_count + + # move core files into session dir + cores = find('core.*', test_subdir) + for core in cores: + dst = core.replace(test_directory, "")[1:] + dst = dst.replace(os.path.sep, "-") + os.rename(core, os.path.join(session_dir, dst)) + + # remove expected timeouts from failures + for xtime in expected_timeout: + if xtime in timed_out: + timed_out.remove(xtime) + failed.remove(xtime) + result = "ExpectedTimeout" + elif xtime in passed: + result = "UnexpectedCompletion" + else: + result = None # failed + + if result: + test_name = os.path.splitext(xtime)[0] + touch(os.path.join(session_dir, "{}-{}".format(result, test_name))) + + # Only run the old summary logic if we don't have a results formatter + # that already prints the summary. + print_legacy_summary = results_formatter is None + if not print_legacy_summary: + # Print summary results. Summarized results at the end always + # get printed to stdout, even if --results-file specifies a different + # file for, say, xUnit output. + results_formatter.print_results(sys.stdout) + + # Figure out exit code by count of test result types. + issue_count = 0 + for issue_status in EventBuilder.TESTRUN_ERROR_STATUS_VALUES: + issue_count += results_formatter.counts_by_test_result_status( + issue_status) + + # Return with appropriate result code + if issue_count > 0: + sys.exit(1) + else: + sys.exit(0) + else: + # Print the legacy test results summary. + print() + sys.stdout.write("Ran %d test suites" % num_test_files) + if num_test_files > 0: + sys.stdout.write(" (%d failed) (%f%%)" % ( + len(failed), 100.0 * len(failed) / num_test_files)) + print() + sys.stdout.write("Ran %d test cases" % num_test_cases) + if num_test_cases > 0: + sys.stdout.write(" (%d failed) (%f%%)" % ( + fail_count, 100.0 * fail_count / num_test_cases)) + print() + exit_code = 0 + + if len(failed) > 0: + failed.sort() + print("Failing Tests (%d)" % len(failed)) + for f in failed: + print("%s: LLDB (suite) :: %s (%s)" % ( + "TIMEOUT" if f in timed_out else "FAIL", f, system_info + )) + exit_code = 1 + + if len(unexpected_successes) > 0: + unexpected_successes.sort() + print("\nUnexpected Successes (%d)" % len(unexpected_successes)) + for u in unexpected_successes: + print("UNEXPECTED SUCCESS: LLDB (suite) :: %s (%s)" % (u, system_info)) + + sys.exit(exit_code) + +if __name__ == '__main__': + sys.stderr.write( + "error: dosep.py no longer supports being called directly. " + "Please call dotest.py directly. The dosep.py-specific arguments " + "have been added under the Parallel processing arguments.\n") + sys.exit(128) |