# coding=utf8 # # (c) Simon Marlow 2002 # import io import shutil import os import re import traceback import time import datetime import copy import glob import sys from math import ceil, trunc from pathlib import Path, PurePath import collections import subprocess from testglobals import config, ghc_env, default_testopts, brokens, t, \ TestRun, TestResult, TestOptions, PerfMetric from testutil import strip_quotes, lndir, link_or_copy_file, passed, \ failBecause, testing_metrics, \ PassFail from term_color import Color, colored import testutil from cpu_features import have_cpu_feature import perf_notes as Perf from perf_notes import MetricChange, PerfStat, MetricOracles extra_src_files = {'T4198': ['exitminus1.c']} # TODO: See #12223 from my_typing import * global pool_sema if config.use_threads: import threading pool_sema = threading.BoundedSemaphore(value=config.threads) global wantToStop wantToStop = False # I have no idea what the type of this is global thisdir_settings thisdir_settings = None # type: ignore def stopNow() -> None: global wantToStop wantToStop = True def stopping() -> bool: return wantToStop _all_ways = None def get_all_ways() -> Set[WayName]: global _all_ways if _all_ways is None: _all_ways = set(config.way_flags.keys()) return _all_ways # Options valid for the current test only (these get reset to # testdir_testopts after each test). global testopts_local if config.use_threads: testopts_local = threading.local() else: class TestOpts_Local: pass testopts_local = TestOpts_Local() # type: ignore def getTestOpts() -> TestOptions: return testopts_local.x def setLocalTestOpts(opts: TestOptions) -> None: global testopts_local testopts_local.x = opts def isCompilerStatsTest() -> bool: opts = getTestOpts() return bool(opts.is_compiler_stats_test) def isStatsTest() -> bool: opts = getTestOpts() return opts.is_stats_test # This can be called at the top of a file of tests, to set default test options # for the following tests. def setTestOpts( f ): global thisdir_settings thisdir_settings = [thisdir_settings, f] # ----------------------------------------------------------------------------- # Canned setup functions for common cases. eg. for a test you might say # # test('test001', normal, compile, ['']) # # to run it without any options, but change it to # # test('test001', expect_fail, compile, ['']) # # to expect failure for this test. # # type TestOpt = (name :: String, opts :: Object) -> IO () def normal( name, opts ): return; def skip( name, opts ): opts.skip = True def expect_fail( name, opts ): # The compiler, testdriver, OS or platform is missing a certain # feature, and we don't plan to or can't fix it now or in the # future. opts.expect = 'fail'; def reqlib( lib ): return lambda name, opts, l=lib: _reqlib (name, opts, l ) def stage1(name, opts): # See Note [Why is there no stage1 setup function?] framework_fail(name, 'stage1 setup function does not exist', 'add your test to testsuite/tests/stage1 instead') # Note [Why is there no stage1 setup function?] # # Presumably a stage1 setup function would signal that the stage1 # compiler should be used to compile a test. # # Trouble is, the path to the compiler + the `ghc --info` settings for # that compiler are currently passed in from the `make` part of the # testsuite driver. # # Switching compilers in the Python part would be entirely too late, as # all ghc_with_* settings would be wrong. See config/ghc for possible # consequences (for example, config.run_ways would still be # based on the default compiler, quite likely causing ./validate --slow # to fail). # # It would be possible to let the Python part of the testsuite driver # make the call to `ghc --info`, but doing so would require quite some # work. Care has to be taken to not affect the run_command tests for # example, as they also use the `ghc --info` settings: # quasiquotation/qq007/Makefile:ifeq "$(GhcDynamic)" "YES" # # If you want a test to run using the stage1 compiler, add it to the # testsuite/tests/stage1 directory. Validate runs the tests in that # directory with `make stage=1`. # Cache the results of looking to see if we have a library or not. # This makes quite a difference, especially on Windows. have_lib_cache = {} # type: Dict[str, bool] def have_library(lib: str) -> bool: """ Test whether the given library is available """ if lib in have_lib_cache: got_it = have_lib_cache[lib] else: cmd = strip_quotes(config.ghc_pkg) p = subprocess.Popen([cmd, '--no-user-package-db', 'describe', lib], stdout=subprocess.PIPE, stderr=subprocess.PIPE, env=ghc_env) # read from stdout and stderr to avoid blocking due to # buffers filling p.communicate() r = p.wait() got_it = r == 0 have_lib_cache[lib] = got_it return got_it def _reqlib( name, opts, lib ): if not have_library(lib): opts.expect = 'missing-lib' def req_haddock( name, opts ): if not config.haddock: opts.expect = 'missing-lib' def req_profiling( name, opts ): '''Require the profiling libraries (add 'GhcLibWays += p' to mk/build.mk)''' if not config.have_profiling: opts.expect = 'fail' def req_shared_libs( name, opts ): if not config.have_shared_libs: opts.expect = 'fail' def req_interp( name, opts ): if not config.have_interp: opts.expect = 'fail' def req_rts_linker( name, opts ): if not config.have_RTS_linker: opts.expect = 'fail' def req_th( name, opts ): """ Mark a test as requiring TemplateHaskell. In addition to having interpreter support, currently this means that we don't run the test in the profasm when when GHC is dynamically-linked since we can't load profiled objects in this case. """ req_interp(name, opts) if ghc_dynamic(): return _omit_ways(name, opts, ['profasm', 'profthreaded']) def req_smp( name, opts ): if not config.have_smp: opts.expect = 'fail' def ignore_stdout(name, opts): opts.ignore_stdout = True def ignore_stderr(name, opts): opts.ignore_stderr = True def combined_output( name, opts ): opts.combined_output = True def use_specs( specs ): """ use_specs allows one to override files based on suffixes. e.g. 'stdout', 'stderr', 'asm', 'prof.sample', etc. Example use_specs({'stdout' : 'prof002.stdout'}) to make the test re-use prof002.stdout. Full Example: test('T5889', [only_ways(['normal']), req_profiling, extra_files(['T5889/A.hs', 'T5889/B.hs']), use_specs({'stdout' : 'prof002.stdout'})], multimod_compile, ['A B', '-O -prof -fno-prof-count-entries -v0']) """ assert isinstance(specs, dict) return lambda name, opts, s=specs: _use_specs( name, opts, s ) def _use_specs( name, opts, specs ): opts.extra_files.extend(specs.values ()) opts.use_specs = specs # ----- def _lint_ways(name: TestName, ways: List[WayName]) -> None: """ Check that all of the ways in a list are valid. """ unknown_ways = [way for way in get_all_ways() if way not in get_all_ways() ] if len(unknown_ways) > 0: framework_fail(name, None, 'Unknown ways: %s' % (unknown_ways,)) def expect_fail_for( ways: List[WayName] ): def helper( name: TestName, opts ): _lint_ways(name, ways) opts.expect_fail_for = ways return helper def expect_broken( bug: IssueNumber ): """ This test is a expected not to work due to the indicated issue number. """ def helper( name: TestName, opts ): record_broken(name, opts, bug) opts.expect = 'fail'; return helper def expect_broken_for( bug: IssueNumber, ways: List[WayName] ): def helper( name: TestName, opts ): _lint_ways(name, ways) record_broken(name, opts, bug) opts.expect_fail_for = ways return helper def record_broken(name: TestName, opts, bug: IssueNumber): me = (bug, opts.testdir, name) if not me in brokens: brokens.append(me) def _expect_pass(way): # Helper function. Not intended for use in .T files. opts = getTestOpts() return opts.expect == 'pass' and way not in opts.expect_fail_for # ----- def fragile( bug: IssueNumber ): """ Indicates that failures of this test should be ignored due to fragility documented in the given ticket. """ def helper( name, opts, bug=bug ): record_broken(name, opts, bug) opts.fragile_ways += config.way_flags.keys() return helper def fragile_for( bug: IssueNumber, ways: List[WayName] ): """ Indicates that failures of this test should be ignored due to fragility in the given test ways as documented in the given ticket. """ def helper( name: TestName, opts ): _lint_ways(name, ways) record_broken(name, opts, bug) opts.fragile_ways += ways return helper # ----- def omit_ways( ways: List[WayName] ): return lambda name, opts: _omit_ways(name, opts, ways) def _omit_ways( name: TestName, opts, ways: List[WayName] ): _lint_ways(name, ways) opts.omit_ways += ways # ----- def only_ways( ways: List[WayName] ): def helper( name: TestName, opts ): _lint_ways(name, ways) opts.only_ways = ways return helper # ----- def extra_ways( ways: List[WayName] ): def helper( name: TestName, opts ): _lint_ways(name, ways) opts.extra_ways = ways return helper # ----- def set_stdin( file ): return lambda name, opts, f=file: _set_stdin(name, opts, f); def _set_stdin( name, opts, f ): opts.stdin = f # ----- def exit_code( val: int ): return lambda name, opts, v=val: _exit_code(name, opts, v); def _exit_code( name, opts, v ): opts.exit_code = v def signal_exit_code( val: int ): if opsys('solaris2'): return exit_code( val ) else: # When application running on Linux receives fatal error # signal, then its exit code is encoded as 128 + signal # value. See http://www.tldp.org/LDP/abs/html/exitcodes.html # I assume that Mac OS X behaves in the same way at least Mac # OS X builder behavior suggests this. return exit_code( val+128 ) # ----- def compile_timeout_multiplier( val: float ): return lambda name, opts, v=val: _compile_timeout_multiplier(name, opts, v) def _compile_timeout_multiplier( name, opts, v ): opts.compile_timeout_multiplier = v def run_timeout_multiplier( val: float ): return lambda name, opts, v=val: _run_timeout_multiplier(name, opts, v) def _run_timeout_multiplier( name, opts, v ): opts.run_timeout_multiplier = v # ----- def extra_run_opts( val ): return lambda name, opts, v=val: _extra_run_opts(name, opts, v); def _extra_run_opts( name, opts, v ): opts.extra_run_opts = v # ----- def extra_hc_opts( val ): return lambda name, opts, v=val: _extra_hc_opts(name, opts, v); def _extra_hc_opts( name, opts, v ): opts.extra_hc_opts = v # ----- def extra_clean( files ): # TODO. Remove all calls to extra_clean. return lambda _name, _opts: None def extra_files(files): return lambda name, opts: _extra_files(name, opts, files) def _extra_files(name, opts, files): opts.extra_files.extend(files) # ----- # Defaults to "test everything, and only break on extreme cases" # # The inputs to this function are slightly interesting: # metric can be either: # - 'all', in which case all 3 possible metrics are collected and compared. # - The specific metric one wants to use in the test. # - A list of the metrics one wants to use in the test. # # Deviation defaults to 20% because the goal is correctness over performance. # The testsuite should avoid breaking when there is not an actual error. # Instead, the testsuite should notify of regressions in a non-breaking manner. # # collect_compiler_stats is used when the metrics collected are about the compiler. # collect_stats is used in the majority case when the metrics to be collected # are about the performance of the runtime code generated by the compiler. def collect_compiler_stats(metric='all',deviation=20): return lambda name, opts, m=metric, d=deviation: _collect_stats(name, opts, m,d, True) def collect_stats(metric='all', deviation=20): return lambda name, opts, m=metric, d=deviation: _collect_stats(name, opts, m, d) # This is an internal function that is used only in the implementation. # 'is_compiler_stats_test' is somewhat of an unfortunate name. # If the boolean is set to true, it indicates that this test is one that # measures the performance numbers of the compiler. # As this is a fairly rare case in the testsuite, it defaults to false to # indicate that it is a 'normal' performance test. def _collect_stats(name: TestName, opts, metrics, deviation, is_compiler_stats_test=False): if not re.match('^[0-9]*[a-zA-Z][a-zA-Z0-9._-]*$', name): failBecause('This test has an invalid name.') # Normalize metrics to a list of strings. if isinstance(metrics, str): if metrics == 'all': metrics = testing_metrics() else: metrics = [metrics] opts.is_stats_test = True if is_compiler_stats_test: opts.is_compiler_stats_test = True tag = 'compile_time' else: tag = 'runtime' # Compiler performance numbers change when debugging is on, making the results # useless and confusing. Therefore, skip if debugging is on. if config.compiler_debugged and is_compiler_stats_test: opts.skip = True for metric_name in metrics: metric = '{}/{}'.format(tag, metric_name) def baselineByWay(way, target_commit, metric=metric): return Perf.baseline_metric( \ target_commit, name, config.test_env, metric, way) opts.stats_range_fields[metric] = MetricOracles(baseline=baselineByWay, deviation=deviation) # ----- def when(b: bool, f): # When list_brokens is on, we want to see all expect_broken calls, # so we always do f if b or config.list_broken: return f else: return normal def unless(b: bool, f): return when(not b, f) def doing_ghci() -> bool: return 'ghci' in config.run_ways def ghc_dynamic() -> bool: return config.ghc_dynamic def fast() -> bool: return config.speed == 2 def platform( plat: str ) -> bool: return config.platform == plat def opsys( os: str ) -> bool: return config.os == os def arch( arch: str ) -> bool: return config.arch == arch def wordsize( ws: int ) -> bool: return config.wordsize == str(ws) def msys( ) -> bool: return config.msys def cygwin( ) -> bool: return config.cygwin def have_vanilla( ) -> bool: return config.have_vanilla def have_ncg( ) -> bool: return config.have_ncg def have_dynamic( ) -> bool: return config.have_dynamic def have_profiling( ) -> bool: return config.have_profiling def in_tree_compiler( ) -> bool: return config.in_tree_compiler def unregisterised( ) -> bool: return config.unregisterised def compiler_profiled( ) -> bool: return config.compiler_profiled def compiler_debugged( ) -> bool: return config.compiler_debugged def have_gdb( ) -> bool: return config.have_gdb def have_readelf( ) -> bool: return config.have_readelf def integer_gmp( ) -> bool: return have_library("integer-gmp") def integer_simple( ) -> bool: return have_library("integer-simple") def llvm_build ( ) -> bool: return config.ghc_built_by_llvm # --- # Note [Measuring residency] # ~~~~~~~~~~~~~~~~~~~~~~~~~~ # # Residency (peak_megabytes_allocated and max_bytes_used) is sensitive # to when the major GC runs, which makes it inherently inaccurate. # Sometime an innocuous change somewhere can shift things around such # that the samples occur at a different time, and the residency # appears to change (up or down) when the underlying profile hasn't # really changed. To further minimize this effect we run with a single # generation (meaning we get a residency sample on every GC) with a small # allocation area (as suggested in #17387). # # However, please don't just ignore changes in residency. If you see # a change in one of these figures, please check whether it is real or # not as follows: # # * Run the test with old and new compilers, adding +RTS -h -i0.01 # (you don't need to compile anything for profiling or enable profiling # libraries to get a heap profile). # * view the heap profiles, read off the maximum residency. If it has # really changed, then you know there's an issue. RESIDENCY_OPTS = '+RTS -A256k -i0 -h -RTS' # See Note [Measuring residency]. def collect_runtime_residency(tolerance_pct: float): return [ collect_stats(['peak_megabytes_allocated', 'max_bytes_used'], tolerance_pct), extra_run_opts(RESIDENCY_OPTS), # The nonmoving collector does not support -G1 omit_ways([WayName(name) for name in ['nonmoving', 'nonmoving_thr', 'nonmoving_thr_ghc']]) ] # See Note [Measuring residency]. def collect_compiler_residency(tolerance_pct: float): return [ collect_compiler_stats(['peak_megabytes_allocated', 'max_bytes_used'], tolerance_pct), extra_hc_opts(RESIDENCY_OPTS), # The nonmoving collector does not support -G1 omit_ways([WayName('nonmoving_thr_ghc')]) ] # --- def high_memory_usage(name, opts): opts.alone = True # If a test is for a multi-CPU race, then running the test alone # increases the chance that we'll actually see it. def multi_cpu_race(name, opts): opts.alone = True # --- def literate( name, opts ): opts.literate = True def c_src( name, opts ): opts.c_src = True def objc_src( name, opts ): opts.objc_src = True def objcpp_src( name, opts ): opts.objcpp_src = True def cmm_src( name, opts ): opts.cmm_src = True def outputdir( odir ): return lambda name, opts, d=odir: _outputdir(name, opts, d) def _outputdir( name, opts, odir ): opts.outputdir = odir; # ---- def pre_cmd( cmd ): return lambda name, opts, c=cmd: _pre_cmd(name, opts, cmd) def _pre_cmd( name, opts, cmd ): opts.pre_cmd = cmd # ---- def cmd_prefix( prefix ): return lambda name, opts, p=prefix: _cmd_prefix(name, opts, prefix) def _cmd_prefix( name, opts, prefix ): opts.cmd_wrapper = lambda cmd, p=prefix: p + ' ' + cmd; # ---- def cmd_wrapper( fun ): return lambda name, opts, f=fun: _cmd_wrapper(name, opts, fun) def _cmd_wrapper( name, opts, fun ): opts.cmd_wrapper = fun # ---- def compile_cmd_prefix( prefix ): return lambda name, opts, p=prefix: _compile_cmd_prefix(name, opts, prefix) def _compile_cmd_prefix( name, opts, prefix ): opts.compile_cmd_prefix = prefix # ---- def check_stdout( f ): return lambda name, opts, f=f: _check_stdout(name, opts, f) def _check_stdout( name, opts, f ): opts.check_stdout = f def no_check_hp(name, opts): opts.check_hp = False # ---- def filter_stdout_lines( regex ): """ Filter lines of stdout with the given regular expression """ def f( name, opts ): _normalise_fun(name, opts, lambda s: '\n'.join(re.findall(regex, s))) return f def normalise_slashes( name, opts ): _normalise_fun(name, opts, normalise_slashes_) def normalise_exe( name, opts ): _normalise_fun(name, opts, normalise_exe_) def normalise_fun( *fs ): return lambda name, opts: _normalise_fun(name, opts, fs) def _normalise_fun( name, opts, *fs ): opts.extra_normaliser = join_normalisers(opts.extra_normaliser, fs) def normalise_errmsg_fun( *fs ): return lambda name, opts: _normalise_errmsg_fun(name, opts, fs) def _normalise_errmsg_fun( name, opts, *fs ): opts.extra_errmsg_normaliser = join_normalisers(opts.extra_errmsg_normaliser, fs) def check_errmsg(needle): def norm(str): if needle in str: return "%s contained in -ddump-simpl\n" % needle else: return "%s not contained in -ddump-simpl\n" % needle return normalise_errmsg_fun(norm) def grep_errmsg(needle): def norm(str): return "".join(filter(lambda l: re.search(needle, l), str.splitlines(True))) return normalise_errmsg_fun(norm) def normalise_whitespace_fun(f): return lambda name, opts: _normalise_whitespace_fun(name, opts, f) def _normalise_whitespace_fun(name, opts, f): opts.whitespace_normaliser = f def normalise_version_( *pkgs ): def normalise_version__( str ): return re.sub('(' + '|'.join(map(re.escape,pkgs)) + ')-[0-9.]+', '\\1-', str) return normalise_version__ def normalise_version( *pkgs ): def normalise_version__( name, opts ): _normalise_fun(name, opts, normalise_version_(*pkgs)) _normalise_errmsg_fun(name, opts, normalise_version_(*pkgs)) return normalise_version__ def normalise_drive_letter(name, opts): # Windows only. Change D:\\ to C:\\. _normalise_fun(name, opts, lambda str: re.sub(r'[A-Z]:\\', r'C:\\', str)) def keep_prof_callstacks(name, opts): """Keep profiling callstacks. Use together with `only_ways(prof_ways)`. """ opts.keep_prof_callstacks = True def join_normalisers(*a): """ Compose functions, flattening sequences. join_normalisers(f1,[f2,f3],f4) is the same as lambda x: f1(f2(f3(f4(x)))) """ def flatten(l): """ Taken from http://stackoverflow.com/a/2158532/946226 """ for el in l: if (isinstance(el, collections.Iterable) and not isinstance(el, (bytes, str))): for sub in flatten(el): yield sub else: yield el a = flatten(a) fn = lambda x:x # identity function for f in a: assert callable(f) fn = lambda x,f=f,fn=fn: fn(f(x)) return fn # ---- # Function for composing two opt-fns together def executeSetups(fs, name, opts): if type(fs) is list: # If we have a list of setups, then execute each one for f in fs: executeSetups(f, name, opts) else: # fs is a single function, so just apply it fs(name, opts) # ----------------------------------------------------------------------------- # The current directory of tests def newTestDir(tempdir, dir): global thisdir_settings # reset the options for this test directory def settings(name, opts, tempdir=tempdir, dir=dir): return _newTestDir(name, opts, tempdir, dir) thisdir_settings = settings # Should be equal to entry in toplevel .gitignore. testdir_suffix = '.run' def _newTestDir(name: TestName, opts: TestOptions, tempdir, dir): testdir = os.path.join('', *(p for p in PurePath(dir).parts if p != '..')) opts.srcdir = Path.cwd() / dir opts.testdir = Path(os.path.join(tempdir, testdir, name + testdir_suffix)) opts.compiler_always_flags = config.compiler_always_flags # ----------------------------------------------------------------------------- # Actually doing tests parallelTests = [] aloneTests = [] allTestNames = set([]) # type: Set[TestName] def runTest(watcher, opts, name: TestName, func, args): if config.use_threads: pool_sema.acquire() t = threading.Thread(target=test_common_thread, name=name, args=(watcher, name, opts, func, args)) t.daemon = False t.start() else: test_common_work(watcher, name, opts, func, args) # name :: String # setup :: [TestOpt] -> IO () def test(name: TestName, setup: "Callable[[List[TestOptions]], None]", func, args) -> None: global aloneTests global parallelTests global allTestNames global thisdir_settings if name in allTestNames: framework_fail(name, None, 'There are multiple tests with this name') if not re.match('^[0-9]*[a-zA-Z][a-zA-Z0-9._-]*$', name): framework_fail(name, None, 'This test has an invalid name') if config.run_only_some_tests: if name not in config.only: return else: # Note [Mutating config.only] # config.only is initially the set of tests requested by # the user (via 'make TEST='). We then remove all tests that # we've already seen (in .T files), so that we can later # report on any tests we couldn't find and error out. config.only.remove(name) # Make a deep copy of the default_testopts, as we need our own copy # of any dictionaries etc inside it. Otherwise, if one test modifies # them, all tests will see the modified version! myTestOpts = copy.deepcopy(default_testopts) executeSetups([thisdir_settings, setup], name, myTestOpts) thisTest = lambda watcher: runTest(watcher, myTestOpts, name, func, args) if myTestOpts.alone: aloneTests.append(thisTest) else: parallelTests.append(thisTest) allTestNames.add(name) if config.use_threads: def test_common_thread(watcher, name, opts, func, args): try: test_common_work(watcher, name, opts, func, args) finally: pool_sema.release() def get_package_cache_timestamp() -> float: if config.package_conf_cache_file is None: return 0.0 else: try: return config.package_conf_cache_file.stat().st_mtime except: return 0.0 do_not_copy = ('.hi', '.o', '.dyn_hi', '.dyn_o', '.out') # 12112 def test_common_work(watcher: testutil.Watcher, name: TestName, opts, func, args) -> None: try: t.total_tests += 1 setLocalTestOpts(opts) package_conf_cache_file_start_timestamp = get_package_cache_timestamp() # All the ways we might run this test if func == compile or func == multimod_compile: all_ways = config.compile_ways elif func == compile_and_run or func == multimod_compile_and_run: all_ways = config.run_ways elif func == ghci_script: if WayName('ghci') in config.run_ways: all_ways = [WayName('ghci')] else: all_ways = [] elif func in [makefile_test, run_command]: # makefile tests aren't necessarily runtime or compile-time # specific. Assume we can run them in all ways. See #16042 for what # happened previously. all_ways = config.compile_ways + config.run_ways else: all_ways = [WayName('normal')] # A test itself can request extra ways by setting opts.extra_ways all_ways = all_ways + [way for way in opts.extra_ways if way not in all_ways] t.total_test_cases += len(all_ways) only_ways = getTestOpts().only_ways ok_way = lambda way: \ not getTestOpts().skip \ and (only_ways is None or (only_ways is not None and way in only_ways)) \ and (config.cmdline_ways == [] or way in config.cmdline_ways) \ and (not (config.skip_perf_tests and isStatsTest())) \ and (not (config.only_perf_tests and not isStatsTest())) \ and way not in getTestOpts().omit_ways # Which ways we are asked to skip do_ways = list(filter (ok_way,all_ways)) # Only run all ways in slow mode. # See Note [validate and testsuite speed] in toplevel Makefile. if config.accept: # Only ever run one way do_ways = do_ways[:1] elif config.speed > 0: # However, if we EXPLICITLY asked for a way (with extra_ways) # please test it! explicit_ways = list(filter(lambda way: way in opts.extra_ways, do_ways)) other_ways = list(filter(lambda way: way not in opts.extra_ways, do_ways)) do_ways = other_ways[:1] + explicit_ways # Find all files in the source directory that this test # depends on. Do this only once for all ways. # Generously add all filenames that start with the name of # the test to this set, as a convenience to test authors. # They will have to use the `extra_files` setup function to # specify all other files that their test depends on (but # this seems to be necessary for only about 10% of all # tests). files = set(f for f in os.listdir(str(opts.srcdir)) if f.startswith(name) and not f == name and not f.endswith(testdir_suffix) and not os.path.splitext(f)[1] in do_not_copy) for filename in (opts.extra_files + extra_src_files.get(name, [])): if filename.startswith('/'): framework_fail(name, None, 'no absolute paths in extra_files please: ' + filename) elif '*' in filename: # Don't use wildcards in extra_files too much, as # globbing is slow. files.update(str(Path(f).relative_to(opts.srcdir)) for f in glob.iglob(str(in_srcdir(filename)))) elif filename: files.add(filename) else: framework_fail(name, None, 'extra_file is empty string') # Run the required tests... for way in do_ways: if stopping(): break try: do_test(name, way, func, args, files) except KeyboardInterrupt: stopNow() except Exception as e: framework_fail(name, way, str(e)) traceback.print_exc() t.n_tests_skipped += len(set(all_ways) - set(do_ways)) if config.cleanup and do_ways: try: cleanup() except Exception as e: framework_fail(name, None, 'Unhandled exception during cleanup: ' + str(e)) package_conf_cache_file_end_timestamp = get_package_cache_timestamp(); if package_conf_cache_file_start_timestamp != package_conf_cache_file_end_timestamp: framework_fail(name, None, 'Package cache timestamps do not match: ' + str(package_conf_cache_file_start_timestamp) + ' ' + str(package_conf_cache_file_end_timestamp)) except Exception as e: framework_fail(name, None, 'Unhandled exception: ' + str(e)) finally: watcher.notify() def do_test(name: TestName, way: WayName, func: Callable[..., PassFail], args, files: Set[str] ) -> None: opts = getTestOpts() full_name = name + '(' + way + ')' progress_args = [ full_name, t.total_tests, len(allTestNames), [len(t.unexpected_passes), len(t.unexpected_failures), len(t.framework_failures)]] if_verbose(2, "=====> {0} {1} of {2} {3}".format(*progress_args)) # Update terminal title # useful progress indicator even when make test VERBOSE=1 if config.supports_colors: print("\033]0;{0} {1} of {2} {3}\007".format(*progress_args), end="") sys.stdout.flush() # Clean up prior to the test, so that we can't spuriously conclude # that it passed on the basis of old run outputs. cleanup() os.makedirs(str(opts.testdir)) # Link all source files for this test into a new directory in # /tmp, and run the test in that directory. This makes it # possible to run tests in parallel, without modification, that # would otherwise (accidentally) write to the same output file. # It also makes it easier to keep the testsuite clean. for extra_file in files: src = in_srcdir(extra_file) dst = in_testdir(os.path.basename(extra_file.rstrip('/\\'))) if src.is_file(): link_or_copy_file(src, dst) elif src.is_dir(): if dst.exists(): shutil.rmtree(str(dst)) dst.mkdir() lndir(src, dst) else: if not config.haddock and os.path.splitext(extra_file)[1] == '.t': # When using a ghc built without haddock support, .t # files are rightfully missing. Don't # framework_fail. Test will be skipped later. pass else: framework_fail(name, way, 'extra_file does not exist: ' + extra_file) if func.__name__ == 'run_command' or func.__name__ == 'makefile_test' or opts.pre_cmd: # When running 'MAKE' make sure 'TOP' still points to the # root of the testsuite. src_makefile = in_srcdir('Makefile') dst_makefile = in_testdir('Makefile') if src_makefile.exists(): makefile = src_makefile.read_text(encoding='UTF-8') makefile = re.sub('TOP=.*', 'TOP=' + config.top, makefile, 1) dst_makefile.write_text(makefile, encoding='UTF-8') if opts.pre_cmd: exit_code = runCmd('cd "{0}" && {1}'.format(opts.testdir, override_options(opts.pre_cmd)), stderr = subprocess.STDOUT, print_output = config.verbose >= 3) # If user used expect_broken then don't record failures of pre_cmd if exit_code != 0 and opts.expect not in ['fail']: framework_fail(name, way, 'pre_cmd failed: {0}'.format(exit_code)) if_verbose(1, '** pre_cmd was "{0}".'.format(override_options(opts.pre_cmd))) result = func(*[name,way] + args) if opts.expect not in ['pass', 'fail', 'missing-lib']: framework_fail(name, way, 'bad expected ' + opts.expect) try: passFail = result.passFail except (KeyError, TypeError): passFail = 'No passFail found' directory = re.sub('^\\.[/\\\\]', '', str(opts.testdir)) if way in opts.fragile_ways: if_verbose(1, '*** fragile test %s resulted in %s' % (full_name, passFail)) if passFail == 'pass': t.fragile_passes.append(TestResult(directory, name, 'fragile', way)) else: t.fragile_failures.append(TestResult(directory, name, 'fragile', way, stdout=result.stdout, stderr=result.stderr)) elif passFail == 'pass': if _expect_pass(way): t.expected_passes.append(TestResult(directory, name, "", way)) t.n_expected_passes += 1 else: if_verbose(1, '*** unexpected pass for %s' % full_name) t.unexpected_passes.append(TestResult(directory, name, 'unexpected', way)) elif passFail == 'fail': if _expect_pass(way): reason = result.reason tag = result.tag if tag == 'stat': if_verbose(1, '*** unexpected stat test failure for %s' % full_name) t.unexpected_stat_failures.append(TestResult(directory, name, reason, way)) else: if_verbose(1, '*** unexpected failure for %s' % full_name) tr = TestResult(directory, name, reason, way, stdout=result.stdout, stderr=result.stderr) t.unexpected_failures.append(tr) else: if opts.expect == 'missing-lib': t.missing_libs.append(TestResult(directory, name, 'missing-lib', way)) else: t.n_expected_failures += 1 else: framework_fail(name, way, 'bad result ' + passFail) # Make is often invoked with -s, which means if it fails, we get # no feedback at all. This is annoying. So let's remove the option # if found and instead have the testsuite decide on what to do # with the output. def override_options(pre_cmd): if config.verbose >= 5 and bool(re.match('\$make', pre_cmd, re.I)): return pre_cmd.replace('-s' , '') \ .replace('--silent', '') \ .replace('--quiet' , '') return pre_cmd def framework_fail(name: Optional[TestName], way: Optional[WayName], reason: str) -> None: opts = getTestOpts() directory = re.sub('^\\.[/\\\\]', '', str(opts.testdir)) full_name = '%s(%s)' % (name, way) if_verbose(1, '*** framework failure for %s %s ' % (full_name, reason)) name2 = name if name is not None else TestName('none') way2 = way if way is not None else WayName('none') t.framework_failures.append(TestResult(directory, name2, reason, way2)) def framework_warn(name: TestName, way: WayName, reason: str) -> None: opts = getTestOpts() directory = re.sub('^\\.[/\\\\]', '', str(opts.testdir)) full_name = name + '(' + way + ')' if_verbose(1, '*** framework warning for %s %s ' % (full_name, reason)) t.framework_warnings.append(TestResult(directory, name, reason, way)) def badResult(result: PassFail) -> bool: try: if result.passFail == 'pass': return False return True except (KeyError, TypeError): return True # ----------------------------------------------------------------------------- # Generic command tests # A generic command test is expected to run and exit successfully. # # The expected exit code can be changed via exit_code() as normal, and # the expected stdout/stderr are stored in .stdout and # .stderr. The output of the command can be ignored # altogether by using the setup function ignore_stdout instead of # run_command. def run_command( name, way, cmd ): return simple_run( name, '', override_options(cmd), '' ) def makefile_test( name, way, target=None ): if target is None: target = name cmd = '$MAKE -s --no-print-directory {target}'.format(target=target) return run_command(name, way, cmd) # ----------------------------------------------------------------------------- # GHCi tests def ghci_script( name, way, script): flags = ' '.join(get_compiler_flags()) way_flags = ' '.join(config.way_flags[way]) # We pass HC and HC_OPTS as environment variables, so that the # script can invoke the correct compiler by using ':! $HC $HC_OPTS' cmd = ('HC={{compiler}} HC_OPTS="{flags}" {{compiler}} {way_flags} {flags}' ).format(flags=flags, way_flags=way_flags) # NB: put way_flags before flags so that flags in all.T can overrie others getTestOpts().stdin = script return simple_run( name, way, cmd, getTestOpts().extra_run_opts ) # ----------------------------------------------------------------------------- # Compile-only tests def compile( name, way, extra_hc_opts ): return do_compile( name, way, False, None, [], extra_hc_opts ) def compile_fail( name, way, extra_hc_opts ): return do_compile( name, way, True, None, [], extra_hc_opts ) def backpack_typecheck( name, way, extra_hc_opts ): return do_compile( name, way, False, None, [], "-fno-code -fwrite-interface " + extra_hc_opts, backpack=True ) def backpack_typecheck_fail( name, way, extra_hc_opts ): return do_compile( name, way, True, None, [], "-fno-code -fwrite-interface " + extra_hc_opts, backpack=True ) def backpack_compile( name, way, extra_hc_opts ): return do_compile( name, way, False, None, [], extra_hc_opts, backpack=True ) def backpack_compile_fail( name, way, extra_hc_opts ): return do_compile( name, way, True, None, [], extra_hc_opts, backpack=True ) def backpack_run( name, way, extra_hc_opts ): return compile_and_run__( name, way, None, [], extra_hc_opts, backpack=True ) def multimod_compile( name, way, top_mod, extra_hc_opts ): return do_compile( name, way, False, top_mod, [], extra_hc_opts ) def multimod_compile_fail( name, way, top_mod, extra_hc_opts ): return do_compile( name, way, True, top_mod, [], extra_hc_opts ) def multimod_compile_filter( name, way, top_mod, extra_hc_opts, filter_with, suppress_stdout=True ): return do_compile( name, way, False, top_mod, [], extra_hc_opts, filter_with=filter_with, suppress_stdout=suppress_stdout ) def multi_compile( name, way, top_mod, extra_mods, extra_hc_opts ): return do_compile( name, way, False, top_mod, extra_mods, extra_hc_opts) def multi_compile_fail( name, way, top_mod, extra_mods, extra_hc_opts ): return do_compile( name, way, True, top_mod, extra_mods, extra_hc_opts) def do_compile(name: TestName, way: WayName, should_fail: bool, top_mod: Optional[Path], extra_mods: List[str], extra_hc_opts: str, **kwargs ) -> PassFail: # print 'Compile only, extra args = ', extra_hc_opts result = extras_build( way, extra_mods, extra_hc_opts ) if badResult(result): return result extra_hc_opts = result.hc_opts result = simple_build(name, way, extra_hc_opts, should_fail, top_mod, False, True, **kwargs) if badResult(result): return result # the actual stderr should always match the expected, regardless # of whether we expected the compilation to fail or not (successful # compilations may generate warnings). expected_stderr_file = find_expected_file(name, 'stderr') actual_stderr_file = add_suffix(name, 'comp.stderr') diff_file_name = in_testdir(add_suffix(name, 'comp.diff')) if not compare_outputs(way, 'stderr', join_normalisers(getTestOpts().extra_errmsg_normaliser, normalise_errmsg), expected_stderr_file, actual_stderr_file, diff_file=diff_file_name, whitespace_normaliser=getattr(getTestOpts(), "whitespace_normaliser", normalise_whitespace)): stderr = diff_file_name.read_text() diff_file_name.unlink() return failBecause('stderr mismatch', stderr=stderr) # no problems found, this test passed return passed() def compile_cmp_asm(name: TestName, way: WayName, ext: str, extra_hc_opts: str ) -> PassFail: print('Compile only, extra args = ', extra_hc_opts) result = simple_build(name + '.' + ext, way, '-keep-s-files -O ' + extra_hc_opts, False, None, False, False) if badResult(result): return result # the actual stderr should always match the expected, regardless # of whether we expected the compilation to fail or not (successful # compilations may generate warnings). expected_asm_file = find_expected_file(name, 'asm') actual_asm_file = add_suffix(name, 's') if not compare_outputs(way, 'asm', join_normalisers(normalise_errmsg, normalise_asm), expected_asm_file, actual_asm_file): return failBecause('asm mismatch') # no problems found, this test passed return passed() def compile_grep_asm(name: TestName, way: WayName, ext: str, is_substring: bool, extra_hc_opts: str ) -> PassFail: print('Compile only, extra args = ', extra_hc_opts) result = simple_build(name + '.' + ext, way, '-keep-s-files -O ' + extra_hc_opts, False, None, False, False) if badResult(result): return result expected_pat_file = find_expected_file(name, 'asm') actual_asm_file = add_suffix(name, 's') if not grep_output(join_normalisers(normalise_errmsg), expected_pat_file, actual_asm_file, is_substring): return failBecause('asm mismatch') # no problems found, this test passed return passed() # ----------------------------------------------------------------------------- # Compile-and-run tests def compile_and_run__(name: TestName, way: WayName, top_mod: Path, extra_mods: List[str], extra_hc_opts: str, backpack: bool=False ) -> PassFail: # print 'Compile and run, extra args = ', extra_hc_opts result = extras_build( way, extra_mods, extra_hc_opts ) if badResult(result): return result extra_hc_opts = result.hc_opts assert extra_hc_opts is not None if way.startswith('ghci'): # interpreted... return interpreter_run(name, way, extra_hc_opts, top_mod) else: # compiled... result = simple_build(name, way, extra_hc_opts, False, top_mod, True, True, backpack = backpack) if badResult(result): return result cmd = './' + name; # we don't check the compiler's stderr for a compile-and-run test return simple_run( name, way, cmd, getTestOpts().extra_run_opts ) def compile_and_run( name, way, extra_hc_opts ): return compile_and_run__( name, way, None, [], extra_hc_opts) def multimod_compile_and_run( name, way, top_mod, extra_hc_opts ): return compile_and_run__( name, way, top_mod, [], extra_hc_opts) def multi_compile_and_run( name, way, top_mod, extra_mods, extra_hc_opts ): return compile_and_run__( name, way, top_mod, extra_mods, extra_hc_opts) def stats( name, way, stats_file ): opts = getTestOpts() return check_stats(name, way, in_testdir(stats_file), opts.stats_range_fields) def static_stats( name, way, stats_file ): opts = getTestOpts() return check_stats(name, way, in_statsdir(stats_file), opts.stats_range_fields) def metric_dict(name, way, metric, value) -> PerfStat: return Perf.PerfStat( test_env = config.test_env, test = name, way = way, metric = metric, value = value) # ----------------------------------------------------------------------------- # Check test stats. This prints the results for the user. # name: name of the test. # way: the way. # stats_file: the path of the stats_file containing the stats for the test. # range_fields: see TestOptions.stats_range_fields # Returns a pass/fail object. Passes if the stats are within the expected value ranges. # This prints the results for the user. def check_stats(name: TestName, way: WayName, stats_file: Path, range_fields: Dict[MetricName, MetricOracles] ) -> PassFail: head_commit = Perf.commit_hash(GitRef('HEAD')) if Perf.inside_git_repo() else None if head_commit is None: return passed() result = passed() if range_fields: try: stats_file_contents = stats_file.read_text() except IOError as e: return failBecause(str(e)) for (metric, baseline_and_dev) in range_fields.items(): # Remove any metric prefix e.g. "runtime/" and "compile_time/" stat_file_metric = metric.split("/")[-1] field_match = re.search('\\("' + stat_file_metric + '", "([0-9]+)"\\)', stats_file_contents) if field_match is None: print('Failed to find metric: ', stat_file_metric) metric_result = failBecause('no such stats metric') else: val = field_match.group(1) assert val is not None actual_val = int(val) # Store the metric so it can later be stored in a git note. perf_stat = metric_dict(name, way, metric, actual_val) change = None # If this is the first time running the benchmark, then pass. baseline = baseline_and_dev.baseline(way, head_commit) \ if Perf.inside_git_repo() else None if baseline is None: metric_result = passed() change = MetricChange.NewMetric else: tolerance_dev = baseline_and_dev.deviation (change, metric_result) = Perf.check_stats_change( perf_stat, baseline, tolerance_dev, config.allowed_perf_changes, config.verbose >= 4) t.metrics.append(PerfMetric(change=change, stat=perf_stat, baseline=baseline)) # If any metric fails then the test fails. # Note, the remaining metrics are still run so that # a complete list of changes can be presented to the user. if metric_result.passFail == 'fail': result = metric_result return result # ----------------------------------------------------------------------------- # Build a single-module program def extras_build( way, extra_mods, extra_hc_opts ): for mod, opts in extra_mods: result = simple_build(mod, way, opts + ' ' + extra_hc_opts, False, None, False, False) if not (mod.endswith('.hs') or mod.endswith('.lhs')): extra_hc_opts += ' %s' % Path(mod).with_suffix('.o') if badResult(result): return result return passed(hc_opts=extra_hc_opts) def simple_build(name: Union[TestName, str], way: WayName, extra_hc_opts: str, should_fail: bool, top_mod: Optional[Path], link: bool, addsuf: bool, backpack: bool = False, suppress_stdout: bool = False, filter_with: str = '') -> Any: opts = getTestOpts() # Redirect stdout and stderr to the same file stdout = in_testdir(name, 'comp.stderr') stderr = subprocess.STDOUT if not suppress_stdout else None if top_mod is not None: srcname = top_mod elif addsuf: if backpack: srcname = add_suffix(name, 'bkp') else: srcname = add_hs_lhs_suffix(name) else: srcname = Path(name) if top_mod is not None: to_do = '--make ' if link: to_do = to_do + '-o ' + name elif backpack: if link: to_do = '-o ' + name + ' ' else: to_do = '' to_do = to_do + '--backpack ' elif link: to_do = '-o ' + name else: to_do = '-c' # just compile stats_file = name + '.comp.stats' if isCompilerStatsTest(): extra_hc_opts += ' +RTS -V0 -t' + stats_file + ' --machine-readable -RTS' if backpack: extra_hc_opts += ' -outputdir ' + name + '.out' # Required by GHC 7.3+, harmless for earlier versions: if (getTestOpts().c_src or getTestOpts().objc_src or getTestOpts().objcpp_src or getTestOpts().cmm_src): extra_hc_opts += ' -no-hs-main ' if getTestOpts().compile_cmd_prefix == '': cmd_prefix = '' else: cmd_prefix = getTestOpts().compile_cmd_prefix + ' ' flags = ' '.join(get_compiler_flags() + config.way_flags[way]) cmd = ('cd "{opts.testdir}" && {cmd_prefix} ' '{{compiler}} {to_do} {srcname} {flags} {extra_hc_opts}' ).format(**locals()) if filter_with != '': cmd = cmd + ' | ' + filter_with exit_code = runCmd(cmd, None, stdout, stderr, opts.compile_timeout_multiplier) actual_stderr_path = in_testdir(name, 'comp.stderr') if exit_code != 0 and not should_fail: if config.verbose >= 1 and _expect_pass(way): print('Compile failed (exit code {0}) errors were:'.format(exit_code)) dump_file(actual_stderr_path) # ToDo: if the sub-shell was killed by ^C, then exit if isCompilerStatsTest(): statsResult = check_stats(TestName(name), way, in_testdir(stats_file), opts.stats_range_fields) if badResult(statsResult): return statsResult if should_fail: if exit_code == 0: stderr_contents = actual_stderr_path.read_text(encoding='UTF-8', errors='replace') return failBecause('exit code 0', stderr=stderr_contents) else: if exit_code != 0: stderr_contents = actual_stderr_path.read_text(encoding='UTF-8', errors='replace') return failBecause('exit code non-0', stderr=stderr_contents) return passed() # ----------------------------------------------------------------------------- # Run a program and check its output # # If testname.stdin exists, route input from that, else # from /dev/null. Route output to testname.run.stdout and # testname.run.stderr. Returns the exit code of the run. def simple_run(name: TestName, way: WayName, prog: str, extra_run_opts: str) -> Any: opts = getTestOpts() # figure out what to use for stdin if opts.stdin: stdin_arg = in_testdir(opts.stdin) # type: Optional[Path] elif in_testdir(name, 'stdin').exists(): stdin_arg = in_testdir(name, 'stdin') else: stdin_arg = None stdout_arg = in_testdir(name, 'run.stdout') if opts.combined_output: stderr_arg = subprocess.STDOUT # type: Union[int, Path] else: stderr_arg = in_testdir(name, 'run.stderr') my_rts_flags = rts_flags(way) # Collect runtime stats if necessary: # isStatsTest and not isCompilerStatsTest(): # assume we are running a ghc compiled program. Collect stats. # isStatsTest and way == 'ghci': # assume we are running a program via ghci. Collect stats stats_file = None # type: Optional[str] if isStatsTest() and (not isCompilerStatsTest() or way == 'ghci'): stats_file = name + '.stats' stats_args = ' +RTS -V0 -t' + stats_file + ' --machine-readable -RTS' else: stats_args = '' # Put extra_run_opts last: extra_run_opts('+RTS foo') should work. cmd = ' '.join([prog, stats_args, my_rts_flags, extra_run_opts]) if opts.cmd_wrapper is not None: cmd = opts.cmd_wrapper(cmd) cmd = 'cd "{opts.testdir}" && {cmd}'.format(**locals()) # run the command exit_code = runCmd(cmd, stdin_arg, stdout_arg, stderr_arg, opts.run_timeout_multiplier) # check the exit code if exit_code != opts.exit_code: if config.verbose >= 1 and _expect_pass(way): print('Wrong exit code for ' + name + '(' + way + ')' + '(expected', opts.exit_code, ', actual', exit_code, ')') dump_stdout(name) dump_stderr(name) return failBecause('bad exit code (%d)' % exit_code) if not (opts.ignore_stderr or stderr_ok(name, way) or opts.combined_output): return failBecause('bad stderr', stderr=read_stderr(name), stdout=read_stdout(name)) if not (opts.ignore_stdout or stdout_ok(name, way)): return failBecause('bad stdout', stderr=read_stderr(name), stdout=read_stdout(name)) check_hp = '-h' in my_rts_flags and opts.check_hp check_prof = '-p' in my_rts_flags # exit_code > 127 probably indicates a crash, so don't try to run hp2ps. if check_hp and (exit_code <= 127 or exit_code == 251) and not check_hp_ok(name): return failBecause('bad heap profile') if check_prof and not check_prof_ok(name, way): return failBecause('bad profile') # Check runtime stats if desired. if stats_file is not None: return check_stats(name, way, in_testdir(stats_file), opts.stats_range_fields) else: return passed() def rts_flags(way: WayName) -> str: args = config.way_rts_flags.get(way, []) return '+RTS {0} -RTS'.format(' '.join(args)) if args else '' # ----------------------------------------------------------------------------- # Run a program in the interpreter and check its output def interpreter_run(name: TestName, way: WayName, extra_hc_opts: str, top_mod: Path ) -> PassFail: opts = getTestOpts() stdout = in_testdir(name, 'interp.stdout') stderr = in_testdir(name, 'interp.stderr') script = in_testdir(name, 'genscript') if opts.combined_output: framework_fail(name, WayName('unsupported'), 'WAY=ghci and combined_output together is not supported') if top_mod is None: srcname = add_hs_lhs_suffix(name) else: srcname = Path(top_mod) delimiter = '===== program output begins here\n' with script.open('w', encoding='UTF-8') as f: # set the prog name and command-line args to match the compiled # environment. f.write(':set prog ' + name + '\n') f.write(':set args ' + opts.extra_run_opts + '\n') # Add marker lines to the stdout and stderr output files, so we # can separate GHCi's output from the program's. f.write(':! echo ' + delimiter) f.write(':! echo 1>&2 ' + delimiter) # Set stdout to be line-buffered to match the compiled environment. f.write('System.IO.hSetBuffering System.IO.stdout System.IO.LineBuffering\n') # wrapping in GHC.TopHandler.runIO ensures we get the same output # in the event of an exception as for the compiled program. f.write('GHC.TopHandler.runIOFastExit Main.main Prelude.>> Prelude.return ()\n') stdin = in_testdir(opts.stdin if opts.stdin else add_suffix(name, 'stdin')) if stdin.exists(): os.system('cat "{0}" >> "{1}"'.format(stdin, script)) flags = ' '.join(get_compiler_flags() + config.way_flags[way]) cmd = ('{{compiler}} {srcname} {flags} {extra_hc_opts}' ).format(**locals()) if opts.cmd_wrapper is not None: cmd = opts.cmd_wrapper(cmd); cmd = 'cd "{opts.testdir}" && {cmd}'.format(**locals()) exit_code = runCmd(cmd, script, stdout, stderr, opts.run_timeout_multiplier) # split the stdout into compilation/program output split_file(stdout, delimiter, in_testdir(name, 'comp.stdout'), in_testdir(name, 'run.stdout')) split_file(stderr, delimiter, in_testdir(name, 'comp.stderr'), in_testdir(name, 'run.stderr')) # check the exit code if exit_code != getTestOpts().exit_code: print('Wrong exit code for ' + name + '(' + way + ') (expected', getTestOpts().exit_code, ', actual', exit_code, ')') dump_stdout(name) dump_stderr(name) return failBecause('bad exit code (%d)' % exit_code, stderr=read_stderr(name), stdout=read_stdout(name)) # ToDo: if the sub-shell was killed by ^C, then exit if not (opts.ignore_stderr or stderr_ok(name, way)): return failBecause('bad stderr', stderr=read_stderr(name), stdout=read_stdout(name)) elif not (opts.ignore_stdout or stdout_ok(name, way)): return failBecause('bad stdout', stderr=read_stderr(name), stdout=read_stdout(name)) else: return passed() def split_file(in_fn: Path, delimiter: str, out1_fn: Path, out2_fn: Path): # See Note [Universal newlines]. with in_fn.open('r', encoding='utf8', errors='replace', newline=None) as infile: with out1_fn.open('w', encoding='utf8', newline='') as out1: with out2_fn.open('w', encoding='utf8', newline='') as out2: line = infile.readline() while re.sub('^\s*','',line) != delimiter and line != '': out1.write(line) line = infile.readline() line = infile.readline() while line != '': out2.write(line) line = infile.readline() # ----------------------------------------------------------------------------- # Utils def get_compiler_flags() -> List[str]: opts = getTestOpts() flags = copy.copy(opts.compiler_always_flags) flags.append(opts.extra_hc_opts) if opts.outputdir is not None: flags.extend(["-outputdir", opts.outputdir]) return flags def stdout_ok(name: TestName, way: WayName) -> bool: actual_stdout_file = add_suffix(name, 'run.stdout') expected_stdout_file = find_expected_file(name, 'stdout') extra_norm = join_normalisers(normalise_output, getTestOpts().extra_normaliser) check_stdout = getTestOpts().check_stdout if check_stdout is not None: actual_stdout_path = in_testdir(actual_stdout_file) return check_stdout(actual_stdout_path, extra_norm) return compare_outputs(way, 'stdout', extra_norm, expected_stdout_file, actual_stdout_file) def read_stdout( name: TestName ) -> str: return in_testdir(name, 'run.stdout').read_text(encoding='UTF-8') def dump_stdout( name: TestName ) -> None: str = read_stdout(name).strip() if str: print("Stdout (", name, "):") print(str) def stderr_ok(name: TestName, way: WayName) -> bool: actual_stderr_file = add_suffix(name, 'run.stderr') expected_stderr_file = find_expected_file(name, 'stderr') return compare_outputs(way, 'stderr', join_normalisers(normalise_errmsg, getTestOpts().extra_errmsg_normaliser), \ expected_stderr_file, actual_stderr_file, whitespace_normaliser=normalise_whitespace) def read_stderr( name: TestName ) -> str: return in_testdir(name, 'run.stderr').read_text(encoding='UTF-8') def dump_stderr( name: TestName ) -> None: str = read_stderr(name).strip() if str: print("Stderr (", name, "):") print(str) def read_no_crs(f: Path) -> str: s = '' try: # See Note [Universal newlines]. with f.open('r', encoding='utf8', errors='replace', newline=None) as h: s = h.read() except Exception: # On Windows, if the program fails very early, it seems the # files stdout/stderr are redirected to may not get created pass return s def write_file(f: Path, s: str) -> None: # See Note [Universal newlines]. with f.open('w', encoding='utf8', newline='') as h: h.write(s) # Note [Universal newlines] # # We don't want to write any Windows style line endings ever, because # it would mean that `make accept` would touch every line of the file # when switching between Linux and Windows. # # Furthermore, when reading a file, it is convenient to translate all # Windows style endings to '\n', as it simplifies searching or massaging # the content. # # Solution: use `Path.open` instead of `open` # * when reading: use newline=None to translate '\r\n' to '\n' # * when writing: use newline='' to not translate '\n' to '\r\n' # # See https://docs.python.org/3/library/pathlib.html#pathlib.Path.open # # This should work with both python2 and python3, and with both mingw* # as msys2 style Python. # # Do note that Path.open returns unicode strings. So we have to specify # the expected encoding. But there is at least one file which is not # valid utf8 (decodingerror002.stdout). Solution: use errors='replace'. # Another solution would be to open files in binary mode always, and # operate on bytes. def check_hp_ok(name: TestName) -> bool: opts = getTestOpts() # do not qualify for hp2ps because we should be in the right directory hp2psCmd = 'cd "{opts.testdir}" && {{hp2ps}} {name}'.format(**locals()) hp2psResult = runCmd(hp2psCmd, print_output=True) actual_ps_path = in_testdir(name, 'ps') if hp2psResult == 0: if actual_ps_path.exists(): if gs_working: gsResult = runCmd(genGSCmd(actual_ps_path)) if (gsResult == 0): return True else: print("hp2ps output for " + name + " is not valid PostScript") return False else: return True # assume postscript is valid without ghostscript else: print("hp2ps did not generate PostScript for " + name) return False else: print("hp2ps error when processing heap profile for " + name) return False def check_prof_ok(name: TestName, way: WayName) -> bool: expected_prof_file = find_expected_file(name, 'prof.sample') expected_prof_path = in_testdir(expected_prof_file) # Check actual prof file only if we have an expected prof file to # compare it with. if not expected_prof_path.exists(): return True actual_prof_file = add_suffix(name, 'prof') actual_prof_path = in_testdir(actual_prof_file) if not actual_prof_path.exists(): print("%s does not exist" % actual_prof_path) return(False) if actual_prof_path.stat().st_size == 0: print("%s is empty" % actual_prof_path) return(False) return compare_outputs(way, 'prof', normalise_prof, expected_prof_file, actual_prof_file, whitespace_normaliser=normalise_whitespace) # Compare expected output to actual output, and optionally accept the # new output. Returns true if output matched or was accepted, false # otherwise. See Note [Output comparison] for the meaning of the # normaliser and whitespace_normaliser parameters. def compare_outputs(way: WayName, kind: str, normaliser: OutputNormalizer, expected_file, actual_file, diff_file=None, whitespace_normaliser: OutputNormalizer=lambda x:x) -> bool: expected_path = in_srcdir(expected_file) actual_path = in_testdir(actual_file) if expected_path.exists(): expected_str = normaliser(read_no_crs(expected_path)) # Create the .normalised file in the testdir, not in the srcdir. expected_normalised_file = add_suffix(expected_file, 'normalised') expected_normalised_path = in_testdir(expected_normalised_file) else: expected_str = '' # See Note [Null device handling] expected_normalised_path = Path(os.devnull) actual_raw = read_no_crs(actual_path) actual_str = normaliser(actual_raw) # See Note [Output comparison]. if whitespace_normaliser(expected_str) == whitespace_normaliser(actual_str): return True else: if config.verbose >= 1 and _expect_pass(way): print('Actual ' + kind + ' output differs from expected:') # See Note [Null device handling] if expected_normalised_path != Path(os.devnull): write_file(expected_normalised_path, expected_str) actual_normalised_path = add_suffix(actual_path, 'normalised') write_file(actual_normalised_path, actual_str) if config.verbose >= 1 and _expect_pass(way): # See Note [Output comparison]. r = runCmd('diff -uw "{0}" "{1}"'.format(null2unix_null(expected_normalised_path), actual_normalised_path), stdout=diff_file, print_output=True) # If for some reason there were no non-whitespace differences, # then do a full diff if r == 0: r = runCmd('diff -u "{0}" "{1}"'.format(null2unix_null(expected_normalised_path), actual_normalised_path), stdout=diff_file, print_output=True) elif diff_file: diff_file.open('ab').close() # Make sure the file exists still as # we will try to read it later if config.accept and (getTestOpts().expect == 'fail' or way in getTestOpts().expect_fail_for): if_verbose(1, 'Test is expected to fail. Not accepting new output.') return False elif config.accept and actual_raw: if config.accept_platform: if_verbose(1, 'Accepting new output for platform "' + config.platform + '".') expected_path += '-' + config.platform elif config.accept_os: if_verbose(1, 'Accepting new output for os "' + config.os + '".') expected_path += '-' + config.os else: if_verbose(1, 'Accepting new output.') write_file(expected_path, actual_raw) return True elif config.accept: if_verbose(1, 'No output. Deleting "{0}".'.format(expected_path)) expected_path.unlink() return True else: return False # Checks that each line from pattern_file is present in actual_file as # a substring or regex pattern depending on is_substring. def grep_output(normaliser: OutputNormalizer, pattern_file, actual_file, is_substring: bool=True): expected_path = in_srcdir(pattern_file) actual_path = in_testdir(actual_file) expected_patterns = read_no_crs(expected_path).strip().split('\n') actual_raw = read_no_crs(actual_path) actual_str = normaliser(actual_raw) success = True failed_patterns = [] def regex_match(pat, actual): return re.search(pat, actual) is not None def substring_match(pat, actual): return pat in actual def is_match(pat, actual): if is_substring: return substring_match(pat, actual) else: return regex_match(pat, actual) for pat in expected_patterns: if not is_match(pat, actual_str): success = False failed_patterns.append(pat) if not success: print('Actual output does not contain the following patterns:') for pat in failed_patterns: print(pat) return success # Note [Output comparison] # # We do two types of output comparison: # # 1. To decide whether a test has failed. We apply a `normaliser` and an # optional `whitespace_normaliser` to the expected and the actual # output, before comparing the two. # # 2. To show as a diff to the user when the test indeed failed. We apply # the same `normaliser` function to the outputs, to make the diff as # small as possible (only showing the actual problem). But we don't # apply the `whitespace_normaliser` here, because it might completely # squash all whitespace, making the diff unreadable. Instead we rely # on the `diff` program to ignore whitespace changes as much as # possible (#10152). # Note [Null device handling] # # On windows the null device is 'nul' instead of '/dev/null'. # This can in principle be easily solved by using os.devnull. # Not doing so causes issues when python tries to read/write/open # the null device. # # However this still leads to a problem when executing shell # commands in the msys environment. Which again expect '/dev/null'. # # So what we do is use os.devnull and convert it to the string # '/dev/null' for shell commands which are bound to run in a # unix-like environment. def null2unix_null(f: Path) -> str: if f == Path(os.devnull): return ('/dev/null') else: return f.as_posix() def normalise_whitespace(s: str) -> str: # Merge contiguous whitespace characters into a single space. return ' '.join(s.split()) callSite_re = re.compile(r', called at (.+):[\d]+:[\d]+ in [\w\-\.]+:') def normalise_callstacks(s: str) -> str: opts = getTestOpts() def repl(matches): location = matches.group(1) location = normalise_slashes_(location) return ', called at {0}:: in :'.format(location) # Ignore line number differences in call stacks (#10834). s = re.sub(callSite_re, repl, s) # Ignore the change in how we identify implicit call-stacks s = s.replace('from ImplicitParams', 'from HasCallStack') if not opts.keep_prof_callstacks: # Don't output prof callstacks. Test output should be # independent from the WAY we run the test. s = re.sub(r'CallStack \(from -prof\):(\n .*)*\n?', '', s) return s tyCon_re = re.compile(r'TyCon\s*\d+L?\#\#\s*\d+L?\#\#\s*', flags=re.MULTILINE) def normalise_type_reps(s: str) -> str: """ Normalise out fingerprints from Typeable TyCon representations """ return re.sub(tyCon_re, 'TyCon FINGERPRINT FINGERPRINT ', s) def normalise_errmsg(s: str) -> str: """Normalise error-messages emitted via stderr""" # IBM AIX's `ld` is a bit chatty if opsys('aix'): s = s.replace('ld: 0706-027 The -x flag is ignored.\n', '') # remove " error:" and lower-case " Warning:" to make patch for # trac issue #10021 smaller s = modify_lines(s, lambda l: re.sub(' error:', '', l)) s = modify_lines(s, lambda l: re.sub(' Warning:', ' warning:', l)) s = normalise_callstacks(s) s = normalise_type_reps(s) # If somefile ends in ".exe" or ".exe:", zap ".exe" (for Windows) # the colon is there because it appears in error messages; this # hacky solution is used in place of more sophisticated filename # mangling s = re.sub('([^\\s])\\.exe', '\\1', s) # normalise slashes, minimise Windows/Unix filename differences s = re.sub('\\\\', '/', s) # The inplace ghc's are called ghc-stage[123] to avoid filename # collisions, so we need to normalise that to just "ghc" s = re.sub('ghc-stage[123]', 'ghc', s) # Error messages sometimes contain integer implementation package s = re.sub('integer-(gmp|simple)-[0-9.]+', 'integer--', s) # Error messages sometimes contain this blurb which can vary # spuriously depending upon build configuration (e.g. based on integer # backend) s = re.sub('...plus ([a-z]+|[0-9]+) instances involving out-of-scope types', '...plus N instances involving out-of-scope types', s) # Also filter out bullet characters. This is because bullets are used to # separate error sections, and tests shouldn't be sensitive to how the # the division happens. bullet = '•'.encode('utf8') if isinstance(s, bytes) else '•' s = s.replace(bullet, '') # Windows only, this is a bug in hsc2hs but it is preventing # stable output for the testsuite. See #9775. For now we filter out this # warning message to get clean output. if config.msys: s = re.sub('Failed to remove file (.*); error= (.*)$', '', s) s = re.sub('DeleteFile "(.+)": permission denied \(Access is denied\.\)(.*)$', '', s) return s # normalise a .prof file, so that we can reasonably compare it against # a sample. This doesn't compare any of the actual profiling data, # only the shape of the profile and the number of entries. def normalise_prof (s: str) -> str: # sip everything up to the line beginning "COST CENTRE" s = re.sub('^(.*\n)*COST CENTRE[^\n]*\n','',s) # sip results for CAFs, these tend to change unpredictably s = re.sub('[ \t]*(CAF|IDLE).*\n','',s) # XXX Ignore Main.main. Sometimes this appears under CAF, and # sometimes under MAIN. s = re.sub('[ \t]*main[ \t]+Main.*\n','',s) # We have something like this: # # MAIN MAIN 53 0 0.0 0.2 0.0 100.0 # CAF Main 105 0 0.0 0.3 0.0 62.5 # readPrec Main Main_1.hs:7:13-16 109 1 0.0 0.6 0.0 0.6 # readPrec Main Main_1.hs:4:13-16 107 1 0.0 0.6 0.0 0.6 # main Main Main_1.hs:(10,1)-(20,20) 106 1 0.0 20.2 0.0 61.0 # == Main Main_1.hs:7:25-26 114 1 0.0 0.0 0.0 0.0 # == Main Main_1.hs:4:25-26 113 1 0.0 0.0 0.0 0.0 # showsPrec Main Main_1.hs:7:19-22 112 2 0.0 1.2 0.0 1.2 # showsPrec Main Main_1.hs:4:19-22 111 2 0.0 0.9 0.0 0.9 # readPrec Main Main_1.hs:7:13-16 110 0 0.0 18.8 0.0 18.8 # readPrec Main Main_1.hs:4:13-16 108 0 0.0 19.9 0.0 19.9 # # then we remove all the specific profiling data, leaving only the cost # centre name, module, src, and entries, to end up with this: (modulo # whitespace between columns) # # MAIN MAIN 0 # readPrec Main Main_1.hs:7:13-16 1 # readPrec Main Main_1.hs:4:13-16 1 # == Main Main_1.hs:7:25-26 1 # == Main Main_1.hs:4:25-26 1 # showsPrec Main Main_1.hs:7:19-22 2 # showsPrec Main Main_1.hs:4:19-22 2 # readPrec Main Main_1.hs:7:13-16 0 # readPrec Main Main_1.hs:4:13-16 0 # Split 9 whitespace-separated groups, take columns 1 (cost-centre), 2 # (module), 3 (src), and 5 (entries). SCC names can't have whitespace, so # this works fine. s = re.sub(r'\s*(\S+)\s*(\S+)\s*(\S+)\s*(\S+)\s*(\S+)\s*(\S+)\s*(\S+)\s*(\S+)\s*(\S+)\s*', '\\1 \\2 \\3 \\5\n', s) return s def normalise_slashes_( s: str ) -> str: s = re.sub('\\\\', '/', s) s = re.sub('//', '/', s) return s def normalise_exe_( s: str ) -> str: s = re.sub('\.exe', '', s) return s def normalise_output( s: str ) -> str: # remove " error:" and lower-case " Warning:" to make patch for # trac issue #10021 smaller s = modify_lines(s, lambda l: re.sub(' error:', '', l)) s = modify_lines(s, lambda l: re.sub(' Warning:', ' warning:', l)) # Remove a .exe extension (for Windows) # This can occur in error messages generated by the program. s = re.sub('([^\\s])\\.exe', '\\1', s) s = normalise_callstacks(s) s = normalise_type_reps(s) return s def normalise_asm( s: str ) -> str: lines = s.split('\n') # Only keep insuctions and labels not starting with a dot. metadata = re.compile('^[ \t]*\\..*$') out = [] for line in lines: # Drop metadata directives (e.g. ".type") if not metadata.match(line): line = re.sub('@plt', '', line) ins = line.lstrip().split() # Drop empty lines. if not ins: continue # Drop operands, except for call insuctions. elif ins[0] == 'call': out.append(ins[0] + ' ' + ins[1]) else: out.append(ins[0]) return '\n'.join(out) def if_verbose( n: int, s: str ) -> None: if config.verbose >= n: print(s) def dump_file(f: Path): try: with f.open() as file: print(file.read()) except Exception: print('') def runCmd(cmd: str, stdin: Union[None, Path]=None, stdout: Union[None, Path]=None, stderr: Union[None, int, Path]=None, timeout_multiplier=1.0, print_output=False) -> int: timeout_prog = strip_quotes(config.timeout_prog) timeout = str(int(ceil(config.timeout * timeout_multiplier))) # Format cmd using config. Example: cmd='{hpc} report A.tix' cmd = cmd.format(**config.__dict__) if_verbose(3, '%s< %s' % (cmd, stdin.name if isinstance(stdin, Path) else '')) stdin_file = stdin.open('rb') if stdin is not None else None stdout_buffer = b'' stderr_buffer = b'' hStdErr = subprocess.PIPE if stderr is subprocess.STDOUT: hStdErr = subprocess.STDOUT try: # cmd is a complex command in Bourne-shell syntax # e.g (cd . && 'C:/users/simonpj/HEAD/inplace/bin/ghc-stage2' ...etc) # Hence it must ultimately be run by a Bourne shell. It's timeout's job # to invoke the Bourne shell r = subprocess.Popen([timeout_prog, timeout, cmd], stdin=stdin_file, stdout=subprocess.PIPE, stderr=hStdErr, env=ghc_env) stdout_buffer, stderr_buffer = r.communicate() finally: if stdin_file: stdin_file.close() if config.verbose >= 1 and print_output: if stdout_buffer: sys.stdout.buffer.write(stdout_buffer) if stderr_buffer: sys.stderr.buffer.write(stderr_buffer) if stdout is not None: if isinstance(stdout, Path): stdout.write_bytes(stdout_buffer) else: with io.open(stdout, 'wb') as f: f.write(stdout_buffer) if stderr is not None: if isinstance(stderr, Path): stderr.write_bytes(stderr_buffer) if r.returncode == 98: # The python timeout program uses 98 to signal that ^C was pressed stopNow() if r.returncode == 99 and getTestOpts().exit_code != 99: # Only print a message when timeout killed the process unexpectedly. if_verbose(1, 'Timeout happened...killed process "{0}"...\n'.format(cmd)) return r.returncode # ----------------------------------------------------------------------------- # checking if ghostscript is available for checking the output of hp2ps def genGSCmd(psfile: Path) -> str: return '{{gs}} -dNODISPLAY -dBATCH -dQUIET -dNOPAUSE "{0}"'.format(psfile) def gsNotWorking() -> None: global gs_working print("GhostScript not available for hp2ps tests") global gs_working gs_working = False if config.have_profiling: if config.gs != '': resultGood = runCmd(genGSCmd(config.top + '/config/good.ps')); if resultGood == 0: resultBad = runCmd(genGSCmd(config.top + '/config/bad.ps') + ' >/dev/null 2>&1') if resultBad != 0: print("GhostScript available for hp2ps tests") gs_working = True else: gsNotWorking(); else: gsNotWorking(); else: gsNotWorking(); def add_suffix( name: Union[str, Path], suffix: str ) -> Path: if suffix == '': return Path(name) else: return Path(str(name) + '.' + suffix) def add_hs_lhs_suffix(name: str) -> Path: if getTestOpts().c_src: return add_suffix(name, 'c') elif getTestOpts().cmm_src: return add_suffix(name, 'cmm') elif getTestOpts().objc_src: return add_suffix(name, 'm') elif getTestOpts().objcpp_src: return add_suffix(name, 'mm') elif getTestOpts().literate: return add_suffix(name, 'lhs') else: return add_suffix(name, 'hs') def in_testdir(name: Union[Path, str], suffix: str='') -> Path: return getTestOpts().testdir / add_suffix(name, suffix) def in_srcdir(name: Union[Path, str], suffix: str='') -> Path: srcdir = getTestOpts().srcdir if srcdir is None: return add_suffix(name, suffix) else: return srcdir / add_suffix(name, suffix) def in_statsdir(name: Union[Path, str], suffix: str='') -> Path: dir = config.stats_files_dir if dir is None: raise TypeError('stats_files_dir is not set') return dir / add_suffix(name, suffix) # Finding the sample output. The filename is of the form # # .stdout[-ws-][-|-] # def find_expected_file(name: TestName, suff: str) -> Path: basename = add_suffix(name, suff) # Override the basename if the user has specified one, this will then be # subjected to the same name mangling scheme as normal to allow platform # specific overrides to work. basename = getTestOpts().use_specs.get(suff, basename) files = [str(basename) + ws + plat for plat in ['-' + config.platform, '-' + config.os, ''] for ws in ['-ws-' + config.wordsize, '']] for f in files: if in_srcdir(f).exists(): return f return basename if config.msys: import stat def cleanup() -> None: testdir = getTestOpts().testdir # type: Path max_attempts = 5 retries = max_attempts def on_error(function, path: str, excinfo): # At least one test (T11489) removes the write bit from a file it # produces. Windows refuses to delete read-only files with a # permission error. Try setting the write bit and try again. Path(path).chmod(stat.S_IWRITE) function(path) # On Windows we have to retry the delete a couple of times. # The reason for this is that a FileDelete command just marks a # file for deletion. The file is really only removed when the last # handle to the file is closed. Unfortunately there are a lot of # system services that can have a file temporarily opened using a shared # readonly lock, such as the built in AV and search indexer. # # We can't really guarantee that these are all off, so what we can do is # whenever after a rmtree the folder still exists to try again and wait a bit. # # Based on what I've seen from the tests on CI server, is that this is relatively rare. # So overall we won't be retrying a lot. If after a reasonable amount of time the folder is # still locked then abort the current test by throwing an exception, this so it won't fail # with an even more cryptic error. # # See #13162 exception = None while retries > 0 and testdir.exists(): time.sleep((max_attempts-retries)*6) try: shutil.rmtree(str(testdir), onerror=on_error, ignore_errors=False) except Exception as e: exception = e retries -= 1 if retries == 0 and testdir.exists(): raise Exception("Unable to remove folder '%s': %s\nUnable to start current test." % (testdir, exception)) else: def cleanup() -> None: testdir = getTestOpts().testdir if testdir.exists(): shutil.rmtree(str(testdir), ignore_errors=False) # ----------------------------------------------------------------------------- # Return a list of all the files ending in '.T' below directories roots. def findTFiles(roots: List[str]) -> Iterator[str]: for root in roots: for path, dirs, files in os.walk(root, topdown=True): # Never pick up .T files in uncleaned .run directories. dirs[:] = [dir for dir in sorted(dirs) if not dir.endswith(testdir_suffix)] for filename in files: if filename.endswith('.T'): yield os.path.join(path, filename) # ----------------------------------------------------------------------------- # Output a test summary to the specified file object def summary(t: TestRun, file: TextIO, short=False, color=False) -> None: file.write('\n') printUnexpectedTests(file, [t.unexpected_passes, t.unexpected_failures, t.unexpected_stat_failures, t.framework_failures]) if short: # Only print the list of unexpected tests above. return if len(t.unexpected_failures) > 0 or \ len(t.unexpected_stat_failures) > 0 or \ len(t.unexpected_passes) > 0 or \ len(t.framework_failures) > 0: summary_color = Color.RED else: summary_color = Color.GREEN assert t.start_time is not None file.write(colored(summary_color, 'SUMMARY') + ' for test run started at ' + t.start_time.strftime("%c %Z") + '\n' + str(datetime.datetime.now() - t.start_time).rjust(8) + ' spent to go through\n' + repr(t.total_tests).rjust(8) + ' total tests, which gave rise to\n' + repr(t.total_test_cases).rjust(8) + ' test cases, of which\n' + repr(t.n_tests_skipped).rjust(8) + ' were skipped\n' + '\n' + repr(len(t.missing_libs)).rjust(8) + ' had missing libraries\n' + repr(t.n_expected_passes).rjust(8) + ' expected passes\n' + repr(t.n_expected_failures).rjust(8) + ' expected failures\n' + '\n' + repr(len(t.framework_failures)).rjust(8) + ' caused framework failures\n' + repr(len(t.framework_warnings)).rjust(8) + ' caused framework warnings\n' + repr(len(t.unexpected_passes)).rjust(8) + ' unexpected passes\n' + repr(len(t.unexpected_failures)).rjust(8) + ' unexpected failures\n' + repr(len(t.unexpected_stat_failures)).rjust(8) + ' unexpected stat failures\n' + repr(len(t.fragile_failures) + len(t.fragile_passes)).rjust(8) + ' fragile tests\n' + '\n') if t.unexpected_passes: file.write('Unexpected passes:\n') printTestInfosSummary(file, t.unexpected_passes) if t.unexpected_failures: file.write('Unexpected failures:\n') printTestInfosSummary(file, t.unexpected_failures) if t.unexpected_stat_failures: file.write('Unexpected stat failures:\n') printTestInfosSummary(file, t.unexpected_stat_failures) if t.framework_failures: file.write('Framework failures:\n') printTestInfosSummary(file, t.framework_failures) if t.framework_warnings: file.write('Framework warnings:\n') printTestInfosSummary(file, t.framework_warnings) if t.fragile_passes: file.write('Fragile test passes:\n') printTestInfosSummary(file, t.fragile_passes) if t.fragile_failures: file.write('Fragile test failures:\n') printTestInfosSummary(file, t.fragile_failures) if stopping(): file.write('WARNING: Testsuite run was terminated early\n') def printUnexpectedTests(file: TextIO, testInfoss): unexpected = set(result.testname for testInfos in testInfoss for result in testInfos if not result.testname.endswith('.T')) if unexpected: file.write('Unexpected results from:\n') file.write('TEST="' + ' '.join(sorted(unexpected)) + '"\n') file.write('\n') def printTestInfosSummary(file: TextIO, testInfos): maxDirLen = max(len(tr.directory) for tr in testInfos) for result in sorted(testInfos, key=lambda r: (r.testname.lower(), r.way, r.directory)): directory = result.directory.ljust(maxDirLen) file.write(' {directory} {r.testname} [{r.reason}] ({r.way})\n'.format( r = result, directory = directory)) file.write('\n') def modify_lines(s: str, f: Callable[[str], str]) -> str: s = '\n'.join([f(l) for l in s.splitlines()]) if s and s[-1] != '\n': # Prevent '\ No newline at end of file' warnings when diffing. s += '\n' return s