/* * Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers * Copyright (c) 1991-1994 by Xerox Corporation. All rights reserved. * Copyright (c) 1996 by Silicon Graphics. All rights reserved. * Copyright (c) 2009-2022 Ivan Maidanski * * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED * OR IMPLIED. ANY USE IS AT YOUR OWN RISK. * * Permission is hereby granted to use or copy this program * for any purpose, provided the above notices are retained on all copies. * Permission to modify the code and to distribute modified code is granted, * provided the above notices are retained, and a notice that the code was * modified is included with the above copyright notice. */ /* An incomplete test for the garbage collector. */ /* Some more obscure entry points are not tested at all. */ /* This must be compiled with the same flags used to build the */ /* GC. It uses GC internals to allow more precise results */ /* checking for some of the tests. */ # ifdef HAVE_CONFIG_H # include "config.h" # endif # undef GC_BUILD #if (defined(DBG_HDRS_ALL) || defined(MAKE_BACK_GRAPH)) \ && !defined(GC_DEBUG) && !defined(CPPCHECK) # define GC_DEBUG #endif #ifdef DEFAULT_VDB /* specified manually (e.g. passed to CFLAGS) */ # define TEST_DEFAULT_VDB #endif #if defined(CPPCHECK) && defined(GC_PTHREADS) && !defined(_GNU_SOURCE) # define _GNU_SOURCE 1 #endif #ifdef GC_NO_THREADS_DISCOVERY # undef GC_NO_THREAD_REDIRECTS #endif #include "gc.h" #include "gc/javaxfc.h" #ifndef NTHREADS /* Number of additional threads to fork. */ # define NTHREADS 5 /* Excludes main thread, which also runs a test. */ /* In the single-threaded case, the number of times to rerun it. */ #endif # if defined(_WIN32_WCE) && !defined(__GNUC__) # include /* # define assert ASSERT */ # else # include /* Not normally used, but handy for debugging. */ # endif #if defined(GC_NO_FINALIZATION) && !defined(NO_TYPED_TEST) # define NO_TYPED_TEST #endif #ifndef NO_TYPED_TEST # include "gc/gc_typed.h" #endif #define NOT_GCBUILD #include "private/gc_priv.h" /* For output, locking, */ /* some statistics and gcconfig.h. */ #if defined(GC_PRINT_VERBOSE_STATS) || defined(GCTEST_PRINT_VERBOSE) # define print_stats VERBOSE # define INIT_PRINT_STATS /* empty */ #else /* Use own variable as GC_print_stats might not be visible. */ static int print_stats = 0; # ifdef GC_READ_ENV_FILE /* GETENV uses GC internal function in this case. */ # define INIT_PRINT_STATS /* empty */ # else # define INIT_PRINT_STATS \ { \ if (0 != GETENV("GC_PRINT_VERBOSE_STATS")) \ print_stats = VERBOSE; \ else if (0 != GETENV("GC_PRINT_STATS")) \ print_stats = 1; \ } # endif #endif /* !GC_PRINT_VERBOSE_STATS */ # ifdef PCR # include "th/PCR_ThCrSec.h" # include "th/PCR_Th.h" # define GC_printf printf # endif # if defined(GC_PTHREADS) && !defined(GC_WIN32_PTHREADS) # include # endif # if ((defined(DARWIN) && defined(MPROTECT_VDB) \ && !defined(MAKE_BACK_GRAPH) && !defined(TEST_HANDLE_FORK)) \ || (defined(THREADS) && !defined(CAN_HANDLE_FORK)) \ || defined(HAVE_NO_FORK) || defined(USE_WINALLOC)) \ && !defined(NO_TEST_HANDLE_FORK) # define NO_TEST_HANDLE_FORK # endif # ifndef NO_TEST_HANDLE_FORK # include # include # include # if defined(HANDLE_FORK) && defined(CAN_CALL_ATFORK) # define INIT_FORK_SUPPORT GC_set_handle_fork(1) /* Causes abort in GC_init on pthread_atfork failure. */ # elif !defined(TEST_FORK_WITHOUT_ATFORK) # define INIT_FORK_SUPPORT GC_set_handle_fork(-1) /* Passing -1 implies fork() should be as well manually */ /* surrounded with GC_atfork_prepare/parent/child. */ # endif # endif # ifndef INIT_FORK_SUPPORT # define INIT_FORK_SUPPORT /* empty */ # endif #ifdef PCR # define FINALIZER_LOCK() PCR_ThCrSec_EnterSys() # define FINALIZER_UNLOCK() PCR_ThCrSec_ExitSys() #elif defined(GC_PTHREADS) static pthread_mutex_t incr_lock = PTHREAD_MUTEX_INITIALIZER; # define FINALIZER_LOCK() pthread_mutex_lock(&incr_lock) # define FINALIZER_UNLOCK() pthread_mutex_unlock(&incr_lock) #elif defined(GC_WIN32_THREADS) static CRITICAL_SECTION incr_cs; # define FINALIZER_LOCK() EnterCriticalSection(&incr_cs) # define FINALIZER_UNLOCK() LeaveCriticalSection(&incr_cs) #else # define FINALIZER_LOCK() (void)0 # define FINALIZER_UNLOCK() (void)0 #endif /* !THREADS */ #include #ifdef TEST_MANUAL_VDB # define INIT_MANUAL_VDB_ALLOWED GC_set_manual_vdb_allowed(1) #elif !defined(SMALL_CONFIG) # define INIT_MANUAL_VDB_ALLOWED GC_set_manual_vdb_allowed(0) #else # define INIT_MANUAL_VDB_ALLOWED /* empty */ #endif #ifdef TEST_PAGES_EXECUTABLE # define INIT_PAGES_EXECUTABLE GC_set_pages_executable(1) #else # define INIT_PAGES_EXECUTABLE (void)0 #endif #define CHECK_GCLIB_VERSION \ if (GC_get_version() != ((GC_VERSION_MAJOR<<16) \ | (GC_VERSION_MINOR<<8) \ | GC_VERSION_MICRO)) { \ GC_printf("libgc version mismatch\n"); \ exit(1); \ } /* Call GC_INIT only on platforms on which we think we really need it, */ /* so that we can test automatic initialization on the rest. */ #if defined(TEST_EXPLICIT_GC_INIT) || defined(AIX) || defined(CYGWIN32) \ || defined(DARWIN) || defined(HOST_ANDROID) \ || (defined(MSWINCE) && !defined(GC_WINMAIN_REDIRECT)) # define GC_OPT_INIT GC_INIT() #else # define GC_OPT_INIT /* empty */ #endif #define INIT_FIND_LEAK \ if (!GC_get_find_leak()) {} else \ GC_printf("This test program is not designed for leak detection mode\n") #ifdef NO_CLOCK # define INIT_PERF_MEASUREMENT (void)0 #else # define INIT_PERF_MEASUREMENT GC_start_performance_measurement() #endif #define GC_COND_INIT() \ INIT_FORK_SUPPORT; INIT_MANUAL_VDB_ALLOWED; INIT_PAGES_EXECUTABLE; \ GC_OPT_INIT; CHECK_GCLIB_VERSION; \ INIT_PRINT_STATS; INIT_FIND_LEAK; INIT_PERF_MEASUREMENT #define CHECK_OUT_OF_MEMORY(p) \ if ((p) == NULL) { \ GC_printf("Out of memory\n"); \ exit(1); \ } /* Define AO primitives for a single-threaded mode. */ #ifndef AO_HAVE_compiler_barrier /* AO_t not defined. */ # define AO_t GC_word #endif #ifndef AO_HAVE_load_acquire static AO_t AO_load_acquire(const volatile AO_t *addr) { AO_t result; FINALIZER_LOCK(); result = *addr; FINALIZER_UNLOCK(); return result; } #endif #ifndef AO_HAVE_store_release /* Not a macro as new_val argument should be evaluated before the lock. */ static void AO_store_release(volatile AO_t *addr, AO_t new_val) { FINALIZER_LOCK(); *addr = new_val; FINALIZER_UNLOCK(); } #endif #ifndef AO_HAVE_fetch_and_add1 # define AO_fetch_and_add1(p) ((*(p))++) /* This is used only to update counters. */ #endif /* Allocation Statistics. Synchronization is not strictly necessary. */ static volatile AO_t uncollectable_count = 0; static volatile AO_t collectable_count = 0; static volatile AO_t atomic_count = 0; static volatile AO_t realloc_count = 0; static volatile AO_t extra_count = 0; /* Amount of space wasted in cons node; */ /* also used in gcj_cons, mktree and */ /* chktree (for other purposes). */ #if defined(GC_AMIGA_FASTALLOC) && defined(AMIGA) EXTERN_C_BEGIN void GC_amiga_free_all_mem(void); EXTERN_C_END void Amiga_Fail(void){GC_amiga_free_all_mem();abort();} # define FAIL Amiga_Fail() #ifndef NO_TYPED_TEST void *GC_amiga_gctest_malloc_explicitly_typed(size_t lb, GC_descr d){ void *ret=GC_malloc_explicitly_typed(lb,d); if(ret==NULL){ GC_gcollect(); ret=GC_malloc_explicitly_typed(lb,d); if(ret==NULL){ GC_printf("Out of memory, (typed allocations are not directly " "supported with the GC_AMIGA_FASTALLOC option.)\n"); FAIL; } } return ret; } void *GC_amiga_gctest_calloc_explicitly_typed(size_t a,size_t lb, GC_descr d){ void *ret=GC_calloc_explicitly_typed(a,lb,d); if(ret==NULL){ GC_gcollect(); ret=GC_calloc_explicitly_typed(a,lb,d); if(ret==NULL){ GC_printf("Out of memory, (typed allocations are not directly " "supported with the GC_AMIGA_FASTALLOC option.)\n"); FAIL; } } return ret; } # define GC_malloc_explicitly_typed(a,b) GC_amiga_gctest_malloc_explicitly_typed(a,b) # define GC_calloc_explicitly_typed(a,b,c) GC_amiga_gctest_calloc_explicitly_typed(a,b,c) #endif /* !NO_TYPED_TEST */ #else /* !AMIGA_FASTALLOC */ # if defined(PCR) || defined(LINT2) # define FAIL abort() # else # define FAIL ABORT("Test failed") # endif #endif /* !AMIGA_FASTALLOC */ /* AT_END may be defined to exercise the interior pointer test */ /* if the collector is configured with ALL_INTERIOR_POINTERS. */ /* As it stands, this test should succeed with either */ /* configuration. In the FIND_LEAK configuration, it should */ /* find lots of leaks, since we free almost nothing. */ struct SEXPR { struct SEXPR * sexpr_car; struct SEXPR * sexpr_cdr; }; typedef struct SEXPR * sexpr; # define INT_TO_SEXPR(x) ((sexpr)(GC_word)(x)) # define SEXPR_TO_INT(x) ((int)(GC_word)(x)) # undef nil # define nil (INT_TO_SEXPR(0)) # define car(x) ((x) -> sexpr_car) # define cdr(x) ((x) -> sexpr_cdr) # define is_nil(x) ((x) == nil) /* Silly implementation of Lisp cons. Intentionally wastes lots of space */ /* to test collector. */ # ifdef VERY_SMALL_CONFIG # define cons small_cons # else sexpr cons (sexpr x, sexpr y) { sexpr r; int *p; unsigned my_extra = (unsigned)AO_fetch_and_add1(&extra_count) % 5000; r = (sexpr)GC_MALLOC(sizeof(struct SEXPR) + my_extra); CHECK_OUT_OF_MEMORY(r); AO_fetch_and_add1(&collectable_count); for (p = (int *)r; (word)p < (word)r + my_extra + sizeof(struct SEXPR); p++) { if (*p) { GC_printf("Found nonzero at %p - allocator is broken\n", (void *)p); FAIL; } *p = (int)((13 << 12) + ((p - (int *)r) & 0xfff)); } # ifdef AT_END r = (sexpr)((char *)r + (my_extra & ~7)); # endif r -> sexpr_car = x; GC_PTR_STORE_AND_DIRTY(&r->sexpr_cdr, y); GC_reachable_here(x); return r; } # endif #include "gc/gc_mark.h" #ifdef GC_GCJ_SUPPORT #include "gc/gc_gcj.h" /* The following struct emulates the vtable in gcj. */ /* This assumes the default value of MARK_DESCR_OFFSET. */ struct fake_vtable { void * dummy; /* class pointer in real gcj. */ GC_word descr; }; struct fake_vtable gcj_class_struct1 = { 0, sizeof(struct SEXPR) + sizeof(struct fake_vtable *) }; /* length based descriptor. */ struct fake_vtable gcj_class_struct2 = { 0, ((GC_word)3 << (CPP_WORDSZ - 3)) | GC_DS_BITMAP}; /* Bitmap based descriptor. */ struct GC_ms_entry * fake_gcj_mark_proc(word * addr, struct GC_ms_entry *mark_stack_ptr, struct GC_ms_entry *mark_stack_limit, word env) { sexpr x; if (1 == env) { /* Object allocated with debug allocator. */ addr = (word *)GC_USR_PTR_FROM_BASE(addr); } x = (sexpr)(addr + 1); /* Skip the vtable pointer. */ mark_stack_ptr = GC_MARK_AND_PUSH( (void *)(x -> sexpr_cdr), mark_stack_ptr, mark_stack_limit, (void * *)&(x -> sexpr_cdr)); mark_stack_ptr = GC_MARK_AND_PUSH( (void *)(x -> sexpr_car), mark_stack_ptr, mark_stack_limit, (void * *)&(x -> sexpr_car)); return mark_stack_ptr; } #endif /* GC_GCJ_SUPPORT */ sexpr small_cons (sexpr x, sexpr y) { sexpr r = GC_NEW(struct SEXPR); CHECK_OUT_OF_MEMORY(r); AO_fetch_and_add1(&collectable_count); r -> sexpr_car = x; GC_PTR_STORE_AND_DIRTY(&r->sexpr_cdr, y); GC_reachable_here(x); return r; } #ifdef NO_CONS_ATOMIC_LEAF # define small_cons_leaf(x) small_cons(INT_TO_SEXPR(x), nil) #else sexpr small_cons_leaf(int x) { sexpr r = (sexpr)GC_MALLOC_ATOMIC(sizeof(struct SEXPR)); CHECK_OUT_OF_MEMORY(r); AO_fetch_and_add1(&atomic_count); r -> sexpr_car = INT_TO_SEXPR(x); r -> sexpr_cdr = nil; return r; } #endif sexpr small_cons_uncollectable (sexpr x, sexpr y) { sexpr r = (sexpr)GC_MALLOC_UNCOLLECTABLE(sizeof(struct SEXPR)); CHECK_OUT_OF_MEMORY(r); AO_fetch_and_add1(&uncollectable_count); r -> sexpr_cdr = (sexpr)(~(GC_word)y); GC_PTR_STORE_AND_DIRTY(&r->sexpr_car, x); return r; } #ifdef GC_GCJ_SUPPORT sexpr gcj_cons(sexpr x, sexpr y) { sexpr result; GC_word cnt = (GC_word)AO_fetch_and_add1(&extra_count); void *d = (cnt & 1) != 0 ? &gcj_class_struct1 : &gcj_class_struct2; size_t lb = sizeof(struct SEXPR) + sizeof(struct fake_vtable*); void *r = (cnt & 2) != 0 ? GC_GCJ_MALLOC_IGNORE_OFF_PAGE(lb + (cnt <= HBLKSIZE / 2 ? cnt : 0), d) : GC_GCJ_MALLOC(lb, d); CHECK_OUT_OF_MEMORY(r); AO_fetch_and_add1(&collectable_count); result = (sexpr)((GC_word *)r + 1); result -> sexpr_car = x; GC_PTR_STORE_AND_DIRTY(&result->sexpr_cdr, y); GC_reachable_here(x); return result; } #endif /* GC_GCJ_SUPPORT */ /* Return reverse(x) concatenated with y */ sexpr reverse1(sexpr x, sexpr y) { if (is_nil(x)) { return y; } else { return reverse1(cdr(x), cons(car(x), y)); } } sexpr reverse(sexpr x) { # ifdef TEST_WITH_SYSTEM_MALLOC GC_noop1(GC_HIDE_POINTER(malloc(100000))); # endif return reverse1(x, nil); } #ifdef GC_PTHREADS /* TODO: Implement for Win32 */ void *do_gcollect(void *arg) { if (print_stats) GC_log_printf("Collect from a standalone thread\n"); GC_gcollect(); return arg; } void collect_from_other_thread(void) { pthread_t t; int code = pthread_create(&t, NULL, do_gcollect, NULL /* arg */); if (code != 0) { GC_printf("gcollect thread creation failed, errno= %d\n", code); FAIL; } code = pthread_join(t, NULL); if (code != 0) { GC_printf("gcollect thread join failed, errno= %d\n", code); FAIL; } } # define MAX_GCOLLECT_THREADS ((NTHREADS+2)/3) static volatile AO_t gcollect_threads_cnt = 0; #endif /* GC_PTHREADS */ sexpr ints(int low, int up) { if (up < 0 ? low > -up : low > up) { if (up < 0) { # ifdef GC_PTHREADS if (AO_fetch_and_add1(&gcollect_threads_cnt) + 1 <= MAX_GCOLLECT_THREADS) { collect_from_other_thread(); return nil; } # endif GC_gcollect_and_unmap(); } return nil; } else { return small_cons(small_cons_leaf(low), ints(low + 1, up)); } } #ifdef GC_GCJ_SUPPORT /* Return reverse(x) concatenated with y */ sexpr gcj_reverse1(sexpr x, sexpr y) { if (is_nil(x)) { return y; } else { return gcj_reverse1(cdr(x), gcj_cons(car(x), y)); } } sexpr gcj_reverse(sexpr x) { return gcj_reverse1(x, nil); } sexpr gcj_ints(int low, int up) { if (low > up) { return nil; } else { return gcj_cons(gcj_cons(INT_TO_SEXPR(low), nil), gcj_ints(low+1, up)); } } #endif /* GC_GCJ_SUPPORT */ /* To check uncollectible allocation we build lists with disguised cdr */ /* pointers, and make sure they don't go away. */ sexpr uncollectable_ints(int low, int up) { if (low > up) { return nil; } else { return small_cons_uncollectable(small_cons_leaf(low), uncollectable_ints(low+1, up)); } } void check_ints(sexpr list, int low, int up) { if (is_nil(list)) { GC_printf("list is nil\n"); FAIL; } if (SEXPR_TO_INT(car(car(list))) != low) { GC_printf( "List reversal produced incorrect list - collector is broken\n"); FAIL; } if (low == up) { if (cdr(list) != nil) { GC_printf("List too long - collector is broken\n"); FAIL; } } else { check_ints(cdr(list), low+1, up); } } # define UNCOLLECTABLE_CDR(x) (sexpr)(~(GC_word)(cdr(x))) void check_uncollectable_ints(sexpr list, int low, int up) { if (SEXPR_TO_INT(car(car(list))) != low) { GC_printf("Uncollectable list corrupted - collector is broken\n"); FAIL; } if (low == up) { if (UNCOLLECTABLE_CDR(list) != nil) { GC_printf("Uncollectable list too long - collector is broken\n"); FAIL; } } else { check_uncollectable_ints(UNCOLLECTABLE_CDR(list), low+1, up); } } /* Not used, but useful for debugging: */ void print_int_list(sexpr x) { if (is_nil(x)) { GC_printf("NIL\n"); } else { GC_printf("(%d)", SEXPR_TO_INT(car(car(x)))); if (!is_nil(cdr(x))) { GC_printf(", "); print_int_list(cdr(x)); } else { GC_printf("\n"); } } } /* ditto: */ void check_marks_int_list(sexpr x) { if (!GC_is_marked(x)) GC_printf("[unm:%p]", (void *)x); else GC_printf("[mkd:%p]", (void *)x); if (is_nil(x)) { GC_printf("NIL\n"); } else { if (!GC_is_marked(car(x))) GC_printf("[unm car:%p]", (void *)car(x)); GC_printf("(%d)", SEXPR_TO_INT(car(car(x)))); if (!is_nil(cdr(x))) { GC_printf(", "); check_marks_int_list(cdr(x)); } else { GC_printf("\n"); } } } /* * A tiny list reversal test to check thread creation. */ #ifdef THREADS # if defined(GC_ENABLE_SUSPEND_THREAD) # include "gc/javaxfc.h" # endif # ifdef VERY_SMALL_CONFIG # define TINY_REVERSE_UPPER_VALUE 4 # else # define TINY_REVERSE_UPPER_VALUE 10 # endif void tiny_reverse_test_inner(void) { int i; for (i = 0; i < 5; ++i) { check_ints(reverse(reverse(ints(1, TINY_REVERSE_UPPER_VALUE))), 1, TINY_REVERSE_UPPER_VALUE); } } # if defined(GC_PTHREADS) void* # elif !defined(MSWINCE) && !defined(MSWIN_XBOX1) && !defined(NO_CRT) \ && !defined(NO_TEST_ENDTHREADEX) # define TEST_ENDTHREADEX unsigned __stdcall # else DWORD __stdcall # endif tiny_reverse_test(void *p_resumed) { # if defined(GC_ENABLE_SUSPEND_THREAD) && !defined(GC_OSF1_THREADS) \ && defined(SIGNAL_BASED_STOP_WORLD) if (p_resumed != NULL) { /* Test self-suspend. */ GC_suspend_thread(pthread_self()); AO_store_release((volatile AO_t *)p_resumed, (AO_t)TRUE); } # else (void)p_resumed; # endif tiny_reverse_test_inner(); # if defined(GC_ENABLE_SUSPEND_THREAD) /* Force collection from a thread. */ GC_gcollect(); # endif # if defined(GC_PTHREADS) && !defined(GC_NO_PTHREAD_CANCEL) { static volatile AO_t tiny_cancel_cnt = 0; if (AO_fetch_and_add1(&tiny_cancel_cnt) % 3 == 0 && GC_pthread_cancel(pthread_self()) != 0) { GC_printf("pthread_cancel failed\n"); FAIL; } } # endif # if defined(GC_PTHREADS) && defined(GC_HAVE_PTHREAD_EXIT) \ || (defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS)) { static volatile AO_t tiny_exit_cnt = 0; if ((AO_fetch_and_add1(&tiny_exit_cnt) & 1) == 0) { # ifdef TEST_ENDTHREADEX GC_endthreadex(0); # elif defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) GC_ExitThread(0); # else GC_pthread_exit(p_resumed); # endif } } # endif return 0; } # if defined(GC_PTHREADS) void fork_a_thread(void) { pthread_t t; int code; # ifdef GC_ENABLE_SUSPEND_THREAD static volatile AO_t forked_cnt = 0; volatile AO_t *p_resumed = NULL; if (AO_fetch_and_add1(&forked_cnt) % 2 == 0) { p_resumed = GC_NEW(AO_t); CHECK_OUT_OF_MEMORY(p_resumed); AO_fetch_and_add1(&collectable_count); } # else # define p_resumed NULL # endif code = pthread_create(&t, NULL, tiny_reverse_test, (void*)p_resumed); if (code != 0) { GC_printf("Small thread creation failed %d\n", code); FAIL; } # if defined(GC_ENABLE_SUSPEND_THREAD) && !defined(GC_OSF1_THREADS) \ && defined(SIGNAL_BASED_STOP_WORLD) if (GC_is_thread_suspended(t) && NULL == p_resumed) { GC_printf("Running thread should be not suspended\n"); FAIL; } GC_suspend_thread(t); /* might be already self-suspended */ if (!GC_is_thread_suspended(t)) { GC_printf("Thread expected to be suspended\n"); FAIL; } GC_suspend_thread(t); /* should be no-op */ for (;;) { GC_resume_thread(t); if (NULL == p_resumed || AO_load_acquire(p_resumed)) break; GC_collect_a_little(); } if (GC_is_thread_suspended(t)) { GC_printf("Resumed thread should be not suspended\n"); FAIL; } GC_resume_thread(t); /* should be no-op */ if (NULL == p_resumed) GC_collect_a_little(); /* Thread could be running or already terminated (but not joined). */ GC_suspend_thread(t); GC_collect_a_little(); if (!GC_is_thread_suspended(t)) { GC_printf("Thread expected to be suspended\n"); FAIL; } GC_resume_thread(t); # endif if ((code = pthread_join(t, 0)) != 0) { GC_printf("Small thread join failed, errno= %d\n", code); FAIL; } } # elif defined(GC_WIN32_THREADS) void fork_a_thread(void) { HANDLE h; # ifdef TEST_ENDTHREADEX unsigned thread_id; h = (HANDLE)GC_beginthreadex(NULL /* security */, 0 /* stack_size */, tiny_reverse_test, NULL /* arglist */, 0 /* initflag */, &thread_id); # else DWORD thread_id; h = CreateThread((SECURITY_ATTRIBUTES *)NULL, (word)0, tiny_reverse_test, NULL, (DWORD)0, &thread_id); /* Explicitly specify types of the */ /* arguments to test the prototype. */ # endif if (h == (HANDLE)NULL) { GC_printf("Small thread creation failed, errcode= %d\n", (int)GetLastError()); FAIL; } if (WaitForSingleObject(h, INFINITE) != WAIT_OBJECT_0) { GC_printf("Small thread wait failed, errcode= %d\n", (int)GetLastError()); FAIL; } } # endif #endif void test_generic_malloc_or_special(void *p) { size_t size; int kind; void *p2; CHECK_OUT_OF_MEMORY(p); kind = GC_get_kind_and_size(p, &size); if (size != GC_size(p)) { GC_printf("GC_get_kind_and_size returned size not matching GC_size\n"); FAIL; } p2 = GC_GENERIC_OR_SPECIAL_MALLOC(10, kind); CHECK_OUT_OF_MEMORY(p2); if (GC_get_kind_and_size(p2, NULL) != kind) { GC_printf("GC_generic_or_special_malloc:" " unexpected kind of returned object\n"); FAIL; } GC_FREE(p2); } /* Try to force a to be strangely aligned */ volatile struct A_s { char dummy; AO_t aa; } A; #define a_set(p) AO_store_release(&A.aa, (AO_t)(p)) #define a_get() (sexpr)AO_load_acquire(&A.aa) /* * Repeatedly reverse lists built out of very different sized cons cells. * Check that we didn't lose anything. */ void *GC_CALLBACK reverse_test_inner(void *data) { int i; sexpr b; sexpr c; sexpr d; sexpr e; sexpr *f, *g, *h; if (data == 0) { /* This stack frame is not guaranteed to be scanned. */ return GC_call_with_gc_active(reverse_test_inner, (void*)(word)1); } # ifndef BIG # if defined(MACOS) \ || (defined(UNIX_LIKE) && defined(NO_GETCONTEXT)) /* e.g. musl */ /* Assume 128 KB stacks at least. */ # if defined(__aarch64__) || defined(__s390x__) # define BIG 600 # else # define BIG 1000 # endif # elif defined(PCR) /* PCR default stack is 100 KB. Stack frames are up to 120 bytes. */ # define BIG 700 # elif defined(MSWINCE) || defined(RTEMS) /* WinCE only allows 64 KB stacks. */ # define BIG 500 # elif defined(EMSCRIPTEN) || defined(OSF1) /* Wasm reports "Maximum call stack size exceeded" error otherwise. */ /* OSF has limited stack space by default, and large frames. */ # define BIG 200 # elif defined(__MACH__) && defined(__ppc64__) # define BIG 2500 # else # define BIG 4500 # endif # endif a_set(ints(1, 49)); b = ints(1, 50); c = ints(1, -BIG); /* force garbage collection inside */ d = uncollectable_ints(1, 100); test_generic_malloc_or_special(d); e = uncollectable_ints(1, 1); /* Check that realloc updates object descriptors correctly */ AO_fetch_and_add1(&collectable_count); f = (sexpr *)GC_MALLOC(4 * sizeof(sexpr)); f = (sexpr *)GC_REALLOC((void *)f, 6 * sizeof(sexpr)); CHECK_OUT_OF_MEMORY(f); AO_fetch_and_add1(&realloc_count); GC_PTR_STORE_AND_DIRTY(f + 5, ints(1, 17)); AO_fetch_and_add1(&collectable_count); g = (sexpr *)GC_MALLOC(513 * sizeof(sexpr)); test_generic_malloc_or_special(g); g = (sexpr *)GC_REALLOC((void *)g, 800 * sizeof(sexpr)); CHECK_OUT_OF_MEMORY(g); AO_fetch_and_add1(&realloc_count); GC_PTR_STORE_AND_DIRTY(g + 799, ints(1, 18)); AO_fetch_and_add1(&collectable_count); h = (sexpr *)GC_MALLOC(1025 * sizeof(sexpr)); h = (sexpr *)GC_REALLOC((void *)h, 2000 * sizeof(sexpr)); CHECK_OUT_OF_MEMORY(h); AO_fetch_and_add1(&realloc_count); # ifdef GC_GCJ_SUPPORT GC_PTR_STORE_AND_DIRTY(h + 1999, gcj_ints(1, 200)); for (i = 0; i < 51; ++i) { GC_PTR_STORE_AND_DIRTY(h + 1999, gcj_reverse(h[1999])); } /* Leave it as the reversed list for now. */ # else GC_PTR_STORE_AND_DIRTY(h + 1999, ints(1, 200)); # endif /* Try to force some collections and reuse of small list elements */ for (i = 0; i < 10; i++) { (void)ints(1, BIG); } /* Superficially test interior pointer recognition on stack */ c = (sexpr)((char *)c + sizeof(char *)); d = (sexpr)((char *)d + sizeof(char *)); GC_FREE((void *)e); check_ints(b,1,50); # ifndef EMSCRIPTEN check_ints(a_get(),1,49); # else /* FIXME: gctest fails unless check_ints(a_get(), ...) are skipped. */ # endif for (i = 0; i < 50; i++) { check_ints(b,1,50); b = reverse(reverse(b)); } check_ints(b,1,50); # ifndef EMSCRIPTEN check_ints(a_get(),1,49); # endif for (i = 0; i < 10 * (NTHREADS+1); i++) { # if (defined(GC_PTHREADS) || defined(GC_WIN32_THREADS)) \ && (NTHREADS > 0) if (i % 10 == 0) fork_a_thread(); # endif /* This maintains the invariant that a always points to a list */ /* of 49 integers. Thus, this is thread safe without locks, */ /* assuming acquire/release barriers in a_get/set() and atomic */ /* pointer assignments (otherwise, e.g., check_ints() may see */ /* an uninitialized object returned by GC_MALLOC). */ a_set(reverse(reverse(a_get()))); # if !defined(AT_END) && !defined(THREADS) /* This is not thread safe, since realloc explicitly deallocates */ a_set(GC_REALLOC(a_get(), (i & 1) != 0 ? 500 : 8200)); AO_fetch_and_add1(&realloc_count); # endif } # ifndef EMSCRIPTEN check_ints(a_get(),1,49); # endif check_ints(b,1,50); /* Restore c and d values. */ c = (sexpr)((char *)c - sizeof(char *)); d = (sexpr)((char *)d - sizeof(char *)); check_ints(c,1,BIG); check_uncollectable_ints(d, 1, 100); check_ints(f[5], 1,17); check_ints(g[799], 1,18); # ifdef GC_GCJ_SUPPORT GC_PTR_STORE_AND_DIRTY(h + 1999, gcj_reverse(h[1999])); # endif check_ints(h[1999], 1,200); # ifndef THREADS a_set(NULL); # endif *(sexpr volatile *)&b = 0; *(sexpr volatile *)&c = 0; return 0; } void reverse_test(void) { /* Test GC_do_blocking/GC_call_with_gc_active. */ (void)GC_do_blocking(reverse_test_inner, 0); } /* * The rest of this builds balanced binary trees, checks that they don't * disappear, and tests finalization. */ typedef struct treenode { int level; struct treenode * lchild; struct treenode * rchild; } tn; #ifndef GC_NO_FINALIZATION int finalizable_count = 0; #endif int finalized_count = 0; int dropped_something = 0; #ifndef GC_NO_FINALIZATION void GC_CALLBACK finalizer(void *obj, void *client_data) { tn *t = (tn *)obj; FINALIZER_LOCK(); if ((int)(GC_word)client_data != t -> level) { GC_printf("Wrong finalization data - collector is broken\n"); FAIL; } finalized_count++; t -> level = -1; /* detect duplicate finalization immediately */ FINALIZER_UNLOCK(); } void GC_CALLBACK dummy_finalizer(void *obj, void *client_data) { UNUSED_ARG(obj); UNUSED_ARG(client_data); } #endif /* !GC_NO_FINALIZATION */ # define MAX_FINALIZED ((NTHREADS+1)*4000) # if !defined(MACOS) GC_FAR GC_word live_indicators[MAX_FINALIZED] = {0}; # ifndef GC_LONG_REFS_NOT_NEEDED GC_FAR void *live_long_refs[MAX_FINALIZED] = { NULL }; # endif #else /* Too big for THINK_C. have to allocate it dynamically. */ GC_word *live_indicators = 0; # ifndef GC_LONG_REFS_NOT_NEEDED # define GC_LONG_REFS_NOT_NEEDED # endif #endif int live_indicators_count = 0; tn * mktree(int n) { tn * result = GC_NEW(tn); tn * left, * right; AO_fetch_and_add1(&collectable_count); # if defined(MACOS) /* get around static data limitations. */ if (!live_indicators) { live_indicators = (GC_word*)NewPtrClear(MAX_FINALIZED * sizeof(GC_word)); CHECK_OUT_OF_MEMORY(live_indicators); } # endif if (0 == n) return NULL; CHECK_OUT_OF_MEMORY(result); result -> level = n; result -> lchild = left = mktree(n - 1); result -> rchild = right = mktree(n - 1); if (AO_fetch_and_add1(&extra_count) % 17 == 0 && n >= 2) { tn * tmp; CHECK_OUT_OF_MEMORY(left); tmp = left -> rchild; CHECK_OUT_OF_MEMORY(right); GC_PTR_STORE_AND_DIRTY(&left->rchild, right->lchild); GC_PTR_STORE_AND_DIRTY(&right->lchild, tmp); } if (AO_fetch_and_add1(&extra_count) % 119 == 0) { # ifndef GC_NO_FINALIZATION int my_index; void **new_link = GC_NEW(void *); CHECK_OUT_OF_MEMORY(new_link); AO_fetch_and_add1(&collectable_count); # endif { FINALIZER_LOCK(); /* Losing a count here causes erroneous report of failure. */ # ifndef GC_NO_FINALIZATION finalizable_count++; my_index = live_indicators_count++; # endif FINALIZER_UNLOCK(); } # ifndef GC_NO_FINALIZATION if (!GC_get_find_leak()) { GC_REGISTER_FINALIZER((void *)result, finalizer, (void *)(GC_word)n, (GC_finalization_proc *)0, (void * *)0); if (my_index >= MAX_FINALIZED) { GC_printf("live_indicators overflowed\n"); FAIL; } live_indicators[my_index] = 13; if (GC_GENERAL_REGISTER_DISAPPEARING_LINK( (void **)(&(live_indicators[my_index])), result) != 0) { GC_printf("GC_general_register_disappearing_link failed\n"); FAIL; } if (GC_move_disappearing_link((void **)(&(live_indicators[my_index])), (void **)(&(live_indicators[my_index]))) != GC_SUCCESS) { GC_printf("GC_move_disappearing_link(link,link) failed\n"); FAIL; } *new_link = (void *)live_indicators[my_index]; if (GC_move_disappearing_link((void **)(&(live_indicators[my_index])), new_link) != GC_SUCCESS) { GC_printf("GC_move_disappearing_link(new_link) failed\n"); FAIL; } /* Note: if other thread is performing fork at this moment, */ /* then the stack of the current thread is dropped (together */ /* with new_link variable) in the child process, and */ /* GC_dl_hashtbl entry with the link equal to new_link will be */ /* removed when a collection occurs (as expected). */ if (GC_unregister_disappearing_link(new_link) == 0) { GC_printf("GC_unregister_disappearing_link failed\n"); FAIL; } if (GC_move_disappearing_link((void **)(&(live_indicators[my_index])), new_link) != GC_NOT_FOUND) { GC_printf("GC_move_disappearing_link(new_link) failed 2\n"); FAIL; } if (GC_GENERAL_REGISTER_DISAPPEARING_LINK( (void **)(&(live_indicators[my_index])), result) != 0) { GC_printf("GC_general_register_disappearing_link failed 2\n"); FAIL; } # ifndef GC_LONG_REFS_NOT_NEEDED if (GC_REGISTER_LONG_LINK(&live_long_refs[my_index], result) != 0) { GC_printf("GC_register_long_link failed\n"); FAIL; } if (GC_move_long_link(&live_long_refs[my_index], &live_long_refs[my_index]) != GC_SUCCESS) { GC_printf("GC_move_long_link(link,link) failed\n"); FAIL; } *new_link = live_long_refs[my_index]; if (GC_move_long_link(&live_long_refs[my_index], new_link) != GC_SUCCESS) { GC_printf("GC_move_long_link(new_link) failed\n"); FAIL; } if (GC_unregister_long_link(new_link) == 0) { GC_printf("GC_unregister_long_link failed\n"); FAIL; } if (GC_move_long_link(&live_long_refs[my_index], new_link) != GC_NOT_FOUND) { GC_printf("GC_move_long_link(new_link) failed 2\n"); FAIL; } if (GC_REGISTER_LONG_LINK(&live_long_refs[my_index], result) != 0) { GC_printf("GC_register_long_link failed 2\n"); FAIL; } # endif } # endif GC_reachable_here(result); } GC_END_STUBBORN_CHANGE(result); GC_reachable_here(left); GC_reachable_here(right); return result; } void chktree(tn *t, int n) { if (0 == n) { if (NULL == t) /* is a leaf? */ return; GC_printf("Clobbered a leaf - collector is broken\n"); FAIL; } if (t -> level != n) { GC_printf("Lost a node at level %d - collector is broken\n", n); FAIL; } if (AO_fetch_and_add1(&extra_count) % 373 == 0) { (void)GC_MALLOC((unsigned)AO_fetch_and_add1(&extra_count) % 5001); AO_fetch_and_add1(&collectable_count); } chktree(t -> lchild, n-1); if (AO_fetch_and_add1(&extra_count) % 73 == 0) { (void)GC_MALLOC((unsigned)AO_fetch_and_add1(&extra_count) % 373); AO_fetch_and_add1(&collectable_count); } chktree(t -> rchild, n-1); } #if defined(GC_PTHREADS) pthread_key_t fl_key; #endif void * alloc8bytes(void) { # ifndef GC_PTHREADS AO_fetch_and_add1(&atomic_count); return GC_MALLOC_ATOMIC(8); # elif defined(SMALL_CONFIG) || defined(GC_DEBUG) AO_fetch_and_add1(&collectable_count); return GC_MALLOC(8); # else void ** my_free_list_ptr; void * my_free_list; void * next; my_free_list_ptr = (void **)pthread_getspecific(fl_key); if (my_free_list_ptr == 0) { my_free_list_ptr = GC_NEW_UNCOLLECTABLE(void *); if (NULL == my_free_list_ptr) return NULL; AO_fetch_and_add1(&uncollectable_count); if (pthread_setspecific(fl_key, my_free_list_ptr) != 0) { GC_printf("pthread_setspecific failed\n"); FAIL; } } my_free_list = *my_free_list_ptr; if (my_free_list == 0) { my_free_list = GC_malloc_many(8); if (NULL == my_free_list) return NULL; } next = GC_NEXT(my_free_list); GC_PTR_STORE_AND_DIRTY(my_free_list_ptr, next); GC_NEXT(my_free_list) = 0; AO_fetch_and_add1(&collectable_count); return my_free_list; # endif } #include "gc/gc_inline.h" void test_tinyfl(void) { void *results[3]; void *tfls[3][GC_TINY_FREELISTS]; # ifndef DONT_ADD_BYTE_AT_END if (GC_get_all_interior_pointers()) return; /* skip */ # endif BZERO(tfls, sizeof(tfls)); /* TODO: Improve testing of FAST_MALLOC functionality. */ GC_MALLOC_WORDS(results[0], 11, tfls[0]); GC_MALLOC_ATOMIC_WORDS(results[1], 20, tfls[1]); GC_CONS(results[2], results[0], results[1], tfls[2]); } void alloc_small(int n) { int i; for (i = 0; i < n; i += 8) { void *p = alloc8bytes(); CHECK_OUT_OF_MEMORY(p); } } # if defined(THREADS) && defined(GC_DEBUG) # ifdef VERY_SMALL_CONFIG # define TREE_HEIGHT 12 # else # define TREE_HEIGHT 15 # endif # else # ifdef VERY_SMALL_CONFIG # define TREE_HEIGHT 13 # else # define TREE_HEIGHT 16 # endif # endif void tree_test(void) { tn * root; int i; root = mktree(TREE_HEIGHT); # ifndef VERY_SMALL_CONFIG alloc_small(5000000); # endif chktree(root, TREE_HEIGHT); FINALIZER_LOCK(); if (finalized_count && !dropped_something) { GC_printf("Premature finalization - collector is broken\n"); FAIL; } dropped_something = 1; FINALIZER_UNLOCK(); GC_reachable_here(root); /* Root needs to remain live until */ /* dropped_something is set. */ root = mktree(TREE_HEIGHT); chktree(root, TREE_HEIGHT); for (i = TREE_HEIGHT; i >= 0; i--) { root = mktree(i); chktree(root, i); } # ifndef VERY_SMALL_CONFIG alloc_small(5000000); # endif } unsigned n_tests = 0; #ifndef NO_TYPED_TEST const GC_word bm_huge[320 / CPP_WORDSZ] = { # if CPP_WORDSZ == 32 0xffffffff, 0xffffffff, 0xffffffff, 0xffffffff, 0xffffffff, # endif (GC_word)((GC_signed_word)-1), (GC_word)((GC_signed_word)-1), (GC_word)((GC_signed_word)-1), (GC_word)((GC_signed_word)-1), ((GC_word)((GC_signed_word)-1)) >> 8 /* highest byte is zero */ }; /* A very simple test of explicitly typed allocation */ void typed_test(void) { GC_word * old, * newP; GC_word bm3[1] = {0}; GC_word bm2[1] = {0}; GC_word bm_large[1] = { 0xf7ff7fff }; GC_descr d1; GC_descr d2; GC_descr d3 = GC_make_descriptor(bm_large, 32); GC_descr d4 = GC_make_descriptor(bm_huge, 320); GC_word * x = (GC_word *)GC_MALLOC_EXPLICITLY_TYPED( 320 * sizeof(GC_word) + 123, d4); # ifndef GC_DEBUG struct GC_calloc_typed_descr_s ctd_l; # endif int i; AO_fetch_and_add1(&collectable_count); (void)GC_make_descriptor(bm_large, 32); if (GC_get_bit(bm_huge, 32) == 0 || GC_get_bit(bm_huge, 311) == 0 || GC_get_bit(bm_huge, 319) != 0) { GC_printf("Bad GC_get_bit() or bm_huge initialization\n"); FAIL; } GC_set_bit(bm3, 0); GC_set_bit(bm3, 1); d1 = GC_make_descriptor(bm3, 2); GC_set_bit(bm2, 1); d2 = GC_make_descriptor(bm2, 2); # ifndef GC_DEBUG if (GC_calloc_prepare_explicitly_typed(&ctd_l, sizeof(ctd_l), 1001, 3 * sizeof(GC_word), d2) != 1) { GC_printf("Out of memory in calloc typed prepare\n"); exit(1); } # endif old = 0; for (i = 0; i < 4000; i++) { if ((i & 0xff) != 0) { newP = (GC_word*)GC_MALLOC_EXPLICITLY_TYPED(4 * sizeof(GC_word), d1); } else { newP = (GC_word*)GC_MALLOC_EXPLICITLY_TYPED_IGNORE_OFF_PAGE( 4 * sizeof(GC_word), d1); } CHECK_OUT_OF_MEMORY(newP); AO_fetch_and_add1(&collectable_count); if (newP[0] != 0 || newP[1] != 0) { GC_printf("Bad initialization by GC_malloc_explicitly_typed\n"); FAIL; } newP[0] = 17; GC_PTR_STORE_AND_DIRTY(newP + 1, old); old = newP; AO_fetch_and_add1(&collectable_count); newP = (GC_word *)GC_MALLOC_EXPLICITLY_TYPED(4 * sizeof(GC_word), d2); CHECK_OUT_OF_MEMORY(newP); newP[0] = 17; GC_PTR_STORE_AND_DIRTY(newP + 1, old); old = newP; AO_fetch_and_add1(&collectable_count); newP = (GC_word*)GC_MALLOC_EXPLICITLY_TYPED(33 * sizeof(GC_word), d3); CHECK_OUT_OF_MEMORY(newP); newP[0] = 17; GC_PTR_STORE_AND_DIRTY(newP + 1, old); old = newP; AO_fetch_and_add1(&collectable_count); newP = (GC_word *)GC_CALLOC_EXPLICITLY_TYPED(4, 2 * sizeof(GC_word), d1); CHECK_OUT_OF_MEMORY(newP); newP[0] = 17; GC_PTR_STORE_AND_DIRTY(newP + 1, old); old = newP; AO_fetch_and_add1(&collectable_count); if (i & 0xff) { newP = (GC_word *)GC_CALLOC_EXPLICITLY_TYPED(7, 3 * sizeof(GC_word), d2); } else { # ifdef GC_DEBUG newP = (GC_word *)GC_CALLOC_EXPLICITLY_TYPED(1001, 3 * sizeof(GC_word), d2); # else newP = (GC_word *)GC_calloc_do_explicitly_typed(&ctd_l, sizeof(ctd_l)); # endif if (newP != NULL && (newP[0] != 0 || newP[1] != 0)) { GC_printf("Bad initialization by GC_calloc_explicitly_typed\n"); FAIL; } } CHECK_OUT_OF_MEMORY(newP); newP[0] = 17; GC_PTR_STORE_AND_DIRTY(newP + 1, old); old = newP; } for (i = 0; i < 20000; i++) { if (newP[0] != 17) { GC_printf("Typed alloc failed at %d\n", i); FAIL; } newP[0] = 0; old = newP; newP = (GC_word *)old[1]; } GC_gcollect(); GC_noop1((word)x); } #endif /* !NO_TYPED_TEST */ #ifdef DBG_HDRS_ALL # define set_print_procs() (void)(A.dummy = 17) #else static volatile AO_t fail_count = 0; void GC_CALLBACK fail_proc1(void *arg) { UNUSED_ARG(arg); AO_fetch_and_add1(&fail_count); } void set_print_procs(void) { /* Set these global variables just once to avoid TSan false positives. */ A.dummy = 17; GC_set_is_valid_displacement_print_proc(fail_proc1); GC_set_is_visible_print_proc(fail_proc1); } # ifdef THREADS # define TEST_FAIL_COUNT(n) 1 # else # define TEST_FAIL_COUNT(n) (fail_count >= (AO_t)(n)) # endif #endif /* !DBG_HDRS_ALL */ static void uniq(void *p, ...) { va_list a; void *q[100]; int n = 0, i, j; q[n++] = p; va_start(a,p); for (;(q[n] = va_arg(a,void *)) != NULL;n++) ; va_end(a); for (i=0; igc_thread_handle, &((struct thr_hndl_sb_s *)cd)->sb); return NULL; } #ifndef MIN_WORDS # define MIN_WORDS 2 #endif void run_one_test(void) { char *x; # ifndef DBG_HDRS_ALL char *y; char **z; # endif # ifndef NO_CLOCK CLOCK_TYPE start_time; CLOCK_TYPE reverse_time; unsigned long time_diff; # endif # ifndef NO_TEST_HANDLE_FORK pid_t pid; int wstatus; # endif struct thr_hndl_sb_s thr_hndl_sb; GC_FREE(0); # ifdef THREADS if (!GC_thread_is_registered() && GC_is_init_called()) { GC_printf("Current thread is not registered with GC\n"); FAIL; } # endif test_tinyfl(); # ifndef DBG_HDRS_ALL AO_fetch_and_add1(&collectable_count); x = (char*)GC_malloc(7); CHECK_OUT_OF_MEMORY(x); AO_fetch_and_add1(&collectable_count); y = (char*)GC_malloc(7); CHECK_OUT_OF_MEMORY(y); if (GC_size(x) != 8 && GC_size(y) != MIN_WORDS * sizeof(GC_word)) { GC_printf("GC_size produced unexpected results\n"); FAIL; } AO_fetch_and_add1(&collectable_count); x = (char*)GC_malloc(15); CHECK_OUT_OF_MEMORY(x); if (GC_size(x) != 16) { GC_printf("GC_size produced unexpected results 2\n"); FAIL; } AO_fetch_and_add1(&collectable_count); x = (char*)GC_malloc(0); CHECK_OUT_OF_MEMORY(x); if (GC_size(x) != MIN_WORDS * sizeof(GC_word)) { GC_printf("GC_malloc(0) failed: GC_size returns %lu\n", (unsigned long)GC_size(x)); FAIL; } AO_fetch_and_add1(&uncollectable_count); x = (char*)GC_malloc_uncollectable(0); CHECK_OUT_OF_MEMORY(x); if (GC_size(x) != MIN_WORDS * sizeof(GC_word)) { GC_printf("GC_malloc_uncollectable(0) failed\n"); FAIL; } AO_fetch_and_add1(&collectable_count); x = (char*)GC_malloc(16); CHECK_OUT_OF_MEMORY(x); if (GC_base(GC_PTR_ADD(x, 13)) != x) { GC_printf("GC_base(heap ptr) produced incorrect result\n"); FAIL; } if (!GC_is_heap_ptr(x)) { GC_printf("GC_is_heap_ptr(heap_ptr) produced incorrect result\n"); FAIL; } if (GC_is_heap_ptr(&x)) { GC_printf("GC_is_heap_ptr(&local_var) produced incorrect result\n"); FAIL; } if (GC_is_heap_ptr((void *)&fail_count) || GC_is_heap_ptr(NULL)) { GC_printf("GC_is_heap_ptr(&global_var) produced incorrect result\n"); FAIL; } (void)GC_PRE_INCR(x, 0); (void)GC_POST_INCR(x); (void)GC_POST_DECR(x); if (GC_base(x) != x) { GC_printf("Bad INCR/DECR result\n"); FAIL; } y = (char *)(GC_word)fail_proc1; # ifndef PCR if (GC_base(y) != 0) { GC_printf("GC_base(fn_ptr) produced incorrect result\n"); FAIL; } # endif if (GC_same_obj(x+5, x) != x + 5) { GC_printf("GC_same_obj produced incorrect result\n"); FAIL; } if (GC_is_visible(y) != y || GC_is_visible(x) != x) { GC_printf("GC_is_visible produced incorrect result\n"); FAIL; } z = (char**)GC_malloc(8); CHECK_OUT_OF_MEMORY(z); AO_fetch_and_add1(&collectable_count); GC_PTR_STORE(z, x); GC_end_stubborn_change(z); if (*z != x) { GC_printf("GC_PTR_STORE failed: %p != %p\n", (void *)(*z), (void *)x); FAIL; } # if !defined(IA64) && !defined(POWERPC) if (!TEST_FAIL_COUNT(1)) { /* On POWERPCs function pointers point to a descriptor in the */ /* data segment, so there should have been no failures. */ /* The same applies to IA64. */ GC_printf("GC_is_visible produced wrong failure indication\n"); FAIL; } # endif if (GC_is_valid_displacement(y) != y || GC_is_valid_displacement(x) != x || GC_is_valid_displacement(x + 3) != x + 3) { GC_printf("GC_is_valid_displacement produced incorrect result\n"); FAIL; } { size_t i; void *p; p = GC_malloc(17); CHECK_OUT_OF_MEMORY(p); AO_fetch_and_add1(&collectable_count); /* TODO: GC_memalign and friends are not tested well. */ for (i = sizeof(GC_word); i <= HBLKSIZE * 4; i *= 2) { p = GC_memalign(i, 17); CHECK_OUT_OF_MEMORY(p); AO_fetch_and_add1(&collectable_count); if ((word)p % i != 0 || *(int *)p != 0) { GC_printf("GC_memalign(%u,17) produced incorrect result: %p\n", (unsigned)i, p); FAIL; } } (void)GC_posix_memalign(&p, 64, 1); AO_fetch_and_add1(&collectable_count); } # ifndef GC_NO_VALLOC { void *p = GC_valloc(78); CHECK_OUT_OF_MEMORY(p); AO_fetch_and_add1(&collectable_count); if (((GC_word)p & 0x1ff /* at least */) != 0 || *(int *)p != 0) { GC_printf("GC_valloc() produced incorrect result: %p\n", p); FAIL; } p = GC_pvalloc(123); CHECK_OUT_OF_MEMORY(p); AO_fetch_and_add1(&collectable_count); /* Note: cannot check GC_size() result. */ if (((GC_word)p & 0x1ff) != 0 || *(int *)p != 0) { GC_printf("GC_pvalloc() produced incorrect result: %p\n", p); FAIL; } } # endif # ifndef ALL_INTERIOR_POINTERS # if defined(POWERPC) if (!TEST_FAIL_COUNT(1)) # else if (!TEST_FAIL_COUNT(GC_get_all_interior_pointers() ? 1 : 2)) # endif { GC_printf( "GC_is_valid_displacement produced wrong failure indication\n"); FAIL; } # endif # endif /* DBG_HDRS_ALL */ x = GC_STRNDUP("abc", 1); CHECK_OUT_OF_MEMORY(x); AO_fetch_and_add1(&atomic_count); if (strlen(x) != 1) { GC_printf("GC_strndup unexpected result\n"); FAIL; } # ifdef GC_REQUIRE_WCSDUP { static const wchar_t ws[] = { 'a', 'b', 'c', 0 }; void *p = GC_WCSDUP(ws); CHECK_OUT_OF_MEMORY(p); AO_fetch_and_add1(&atomic_count); } # endif # ifndef GC_NO_FINALIZATION if (!GC_get_find_leak()) { void **p = (void **)GC_MALLOC_ATOMIC(sizeof(void*)); CHECK_OUT_OF_MEMORY(p); AO_fetch_and_add1(&atomic_count); *p = x; if (GC_register_disappearing_link(p) != 0) { GC_printf("GC_register_disappearing_link failed\n"); FAIL; } if (GC_get_java_finalization()) { GC_finalization_proc ofn = 0; void *ocd = NULL; GC_REGISTER_FINALIZER_UNREACHABLE(p, dummy_finalizer, NULL, &ofn, &ocd); if (ofn != 0 || ocd != NULL) { GC_printf("GC_register_finalizer_unreachable unexpected result\n"); FAIL; } } # ifndef GC_TOGGLE_REFS_NOT_NEEDED (void)GC_toggleref_add(p, 1); # endif } # endif /* Test floating point alignment */ { double *dp = GC_NEW(double); CHECK_OUT_OF_MEMORY(dp); AO_fetch_and_add1(&collectable_count); *dp = 1.0; dp = GC_NEW(double); CHECK_OUT_OF_MEMORY(dp); AO_fetch_and_add1(&collectable_count); *dp = 1.0; # ifndef NO_DEBUGGING (void)GC_count_set_marks_in_hblk(dp); # endif } /* Test size 0 allocation a bit more */ { size_t i; for (i = 0; i < 10000; ++i) { (void)GC_MALLOC(0); AO_fetch_and_add1(&collectable_count); GC_FREE(GC_MALLOC(0)); (void)GC_MALLOC_ATOMIC(0); AO_fetch_and_add1(&atomic_count); GC_FREE(GC_MALLOC_ATOMIC(0)); test_generic_malloc_or_special(GC_malloc_atomic(1)); AO_fetch_and_add1(&atomic_count); GC_FREE(GC_MALLOC_ATOMIC_IGNORE_OFF_PAGE(1)); GC_FREE(GC_MALLOC_IGNORE_OFF_PAGE(2)); (void)GC_generic_malloc_ignore_off_page(2 * HBLKSIZE, NORMAL); AO_fetch_and_add1(&collectable_count); } } thr_hndl_sb.gc_thread_handle = GC_get_my_stackbottom(&thr_hndl_sb.sb); # ifdef GC_GCJ_SUPPORT GC_REGISTER_DISPLACEMENT(sizeof(struct fake_vtable *)); GC_init_gcj_malloc(0, (void *)(GC_word)fake_gcj_mark_proc); # endif /* Make sure that fn arguments are visible to the collector. */ uniq( GC_malloc(12), GC_malloc(12), GC_malloc(12), (GC_gcollect(),GC_malloc(12)), GC_malloc(12), GC_malloc(12), GC_malloc(12), (GC_gcollect(),GC_malloc(12)), GC_malloc(12), GC_malloc(12), GC_malloc(12), (GC_gcollect(),GC_malloc(12)), GC_malloc(12), GC_malloc(12), GC_malloc(12), (GC_gcollect(),GC_malloc(12)), GC_malloc(12), GC_malloc(12), GC_malloc(12), (GC_gcollect(),GC_malloc(12)), (void *)0); /* GC_malloc(0) must return NULL or something we can deallocate. */ GC_free(GC_malloc(0)); GC_free(GC_malloc_atomic(0)); GC_free(GC_malloc(0)); GC_free(GC_malloc_atomic(0)); # ifndef NO_TEST_HANDLE_FORK GC_atfork_prepare(); pid = fork(); if (pid != 0) { GC_atfork_parent(); if (pid == -1) { GC_printf("Process fork failed\n"); FAIL; } if (print_stats) GC_log_printf("Forked child process, pid= %ld\n", (long)pid); if (waitpid(pid, &wstatus, 0) == -1) { GC_printf("Wait for child process failed\n"); FAIL; } if (!WIFEXITED(wstatus) || WEXITSTATUS(wstatus) != 0) { GC_printf("Child process failed, pid= %ld, status= 0x%x\n", (long)pid, wstatus); FAIL; } } else { pid_t child_pid = getpid(); GC_atfork_child(); if (print_stats) GC_log_printf("Started a child process, pid= %ld\n", (long)child_pid); # ifdef PARALLEL_MARK GC_gcollect(); /* no parallel markers */ # endif GC_start_mark_threads(); GC_gcollect(); # ifdef THREADS /* Skip "Premature finalization" check in the */ /* child process because there could be a chance */ /* that some other thread of the parent was */ /* executing mktree at the moment of fork. */ dropped_something = 1; # endif tree_test(); # if !defined(DBG_HDRS_ALL) && !defined(NO_TYPED_TEST) typed_test(); # endif # ifdef THREADS if (print_stats) GC_log_printf("Starting tiny reverse test, pid= %ld\n", (long)child_pid); tiny_reverse_test_inner(); GC_gcollect(); # endif if (print_stats) GC_log_printf("Finished a child process, pid= %ld\n", (long)child_pid); exit(0); } # endif (void)GC_call_with_alloc_lock(set_stackbottom, &thr_hndl_sb); /* Repeated list reversal test. */ # ifndef NO_CLOCK GET_TIME(start_time); # endif reverse_test(); # ifndef NO_CLOCK if (print_stats) { GET_TIME(reverse_time); time_diff = MS_TIME_DIFF(reverse_time, start_time); GC_log_printf("Finished reverse_test at time %u (%p)\n", (unsigned) time_diff, (void *)&start_time); } # endif # if !defined(DBG_HDRS_ALL) && !defined(NO_TYPED_TEST) typed_test(); # ifndef NO_CLOCK if (print_stats) { CLOCK_TYPE typed_time; GET_TIME(typed_time); time_diff = MS_TIME_DIFF(typed_time, start_time); GC_log_printf("Finished typed_test at time %u (%p)\n", (unsigned) time_diff, (void *)&start_time); } # endif # endif /* DBG_HDRS_ALL */ tree_test(); # ifdef TEST_WITH_SYSTEM_MALLOC free(calloc(1,1)); free(realloc(NULL, 64)); # endif # ifndef NO_CLOCK if (print_stats) { CLOCK_TYPE tree_time; GET_TIME(tree_time); time_diff = MS_TIME_DIFF(tree_time, start_time); GC_log_printf("Finished tree_test at time %u (%p)\n", (unsigned) time_diff, (void *)&start_time); } # endif /* Run reverse_test a second time, so we hopefully notice corruption. */ reverse_test(); # ifndef NO_DEBUGGING (void)GC_is_tmp_root((/* no volatile */ void *)&atomic_count); # endif # ifndef NO_CLOCK if (print_stats) { GET_TIME(reverse_time); time_diff = MS_TIME_DIFF(reverse_time, start_time); GC_log_printf("Finished second reverse_test at time %u (%p)\n", (unsigned)time_diff, (void *)&start_time); } # endif /* GC_allocate_ml and GC_need_to_lock are no longer exported, and */ /* AO_fetch_and_add1() may be unavailable to update a counter. */ (void)GC_call_with_alloc_lock(inc_int_counter, &n_tests); # ifndef NO_CLOCK if (print_stats) GC_log_printf("Finished %p\n", (void *)&start_time); # endif } /* Execute some tests after termination of other test threads (if any). */ void run_single_threaded_test(void) { GC_disable(); GC_FREE(GC_MALLOC(100)); GC_expand_hp(0); /* add a block to heap */ GC_enable(); } void GC_CALLBACK reachable_objs_counter(void *obj, size_t size, void *pcounter) { if (0 == size) { GC_printf("Reachable object has zero size\n"); FAIL; } if (GC_base(obj) != obj) { GC_printf("Invalid reachable object base passed by enumerator: %p\n", obj); FAIL; } if (GC_size(obj) != size) { GC_printf("Invalid reachable object size passed by enumerator: %lu\n", (unsigned long)size); FAIL; } (*(unsigned *)pcounter)++; } #define NUMBER_ROUND_UP(v, bound) ((((v) + (bound) - 1) / (bound)) * (bound)) void check_heap_stats(void) { size_t max_heap_sz; int i; # ifndef GC_NO_FINALIZATION # ifdef FINALIZE_ON_DEMAND int late_finalize_count = 0; # endif # endif unsigned obj_count = 0; if (!GC_is_init_called()) { GC_printf("GC should be initialized!\n"); FAIL; } # ifdef VERY_SMALL_CONFIG /* The upper bounds are a guess, which has been empirically */ /* adjusted. On low end uniprocessors with incremental GC */ /* these may be particularly dubious, since empirically the */ /* heap tends to grow largely as a result of the GC not */ /* getting enough cycles. */ # if CPP_WORDSZ == 64 max_heap_sz = 4500000; # else max_heap_sz = 2800000; # endif # else # if CPP_WORDSZ == 64 max_heap_sz = 26000000; # else max_heap_sz = 16000000; # endif # endif # ifdef GC_DEBUG max_heap_sz *= 2; # ifdef SAVE_CALL_CHAIN max_heap_sz *= 3; # ifdef SAVE_CALL_COUNT max_heap_sz += max_heap_sz * NFRAMES / 4; # endif # endif # endif # if defined(ADDRESS_SANITIZER) && !defined(__clang__) max_heap_sz = max_heap_sz * 2 - max_heap_sz / 3; # endif # ifdef MEMORY_SANITIZER max_heap_sz += max_heap_sz / 4; # endif max_heap_sz *= n_tests; # if defined(USE_MMAP) || defined(MSWIN32) max_heap_sz = NUMBER_ROUND_UP(max_heap_sz, 4 * 1024 * 1024); # endif /* Garbage collect repeatedly so that all inaccessible objects */ /* can be finalized. */ while (GC_collect_a_little()) { } /* should work even if disabled GC */ for (i = 0; i < 16; i++) { GC_gcollect(); # ifndef GC_NO_FINALIZATION # ifdef FINALIZE_ON_DEMAND late_finalize_count += # endif GC_invoke_finalizers(); # endif } if (print_stats) { struct GC_stack_base sb; int res = GC_get_stack_base(&sb); if (res == GC_SUCCESS) { GC_log_printf("Primordial thread stack bottom: %p\n", sb.mem_base); } else if (res == GC_UNIMPLEMENTED) { GC_log_printf("GC_get_stack_base() unimplemented\n"); } else { GC_printf("GC_get_stack_base() failed: %d\n", res); FAIL; } } GC_alloc_lock(); GC_enumerate_reachable_objects_inner(reachable_objs_counter, &obj_count); GC_alloc_unlock(); GC_printf("Completed %u tests\n", n_tests); GC_printf("Allocated %d collectable objects\n", (int)collectable_count); GC_printf("Allocated %d uncollectable objects\n", (int)uncollectable_count); GC_printf("Allocated %d atomic objects\n", (int)atomic_count); GC_printf("Reallocated %d objects\n", (int)realloc_count); # ifndef NO_TEST_HANDLE_FORK GC_printf("Garbage collection after fork is tested too\n"); # endif # ifndef GC_NO_FINALIZATION if (!GC_get_find_leak()) { int still_live = 0; # ifndef GC_LONG_REFS_NOT_NEEDED int still_long_live = 0; # endif # ifdef FINALIZE_ON_DEMAND if (finalized_count != late_finalize_count) { GC_printf("Finalized %d/%d objects - demand finalization error\n", finalized_count, finalizable_count); FAIL; } # endif if (finalized_count > finalizable_count || finalized_count < finalizable_count/2) { GC_printf("Finalized %d/%d objects - " "finalization is probably broken\n", finalized_count, finalizable_count); FAIL; } else { GC_printf("Finalized %d/%d objects - finalization is probably OK\n", finalized_count, finalizable_count); } for (i = 0; i < MAX_FINALIZED; i++) { if (live_indicators[i] != 0) { still_live++; } # ifndef GC_LONG_REFS_NOT_NEEDED if (live_long_refs[i] != NULL) { still_long_live++; } # endif } i = finalizable_count - finalized_count - still_live; if (0 != i) { GC_printf("%d disappearing links remain and %d more objects " "were not finalized\n", still_live, i); if (i > 10) { GC_printf("\tVery suspicious!\n"); } else { GC_printf("\tSlightly suspicious, but probably OK\n"); } } # ifndef GC_LONG_REFS_NOT_NEEDED if (0 != still_long_live) { GC_printf("%d 'long' links remain\n", still_long_live); } # endif } # endif GC_printf("Total number of bytes allocated is %lu\n", (unsigned long)GC_get_total_bytes()); GC_printf("Total memory use by allocated blocks is %lu bytes\n", (unsigned long)GC_get_memory_use()); GC_printf("Final heap size is %lu bytes\n", (unsigned long)GC_get_heap_size()); if (GC_get_total_bytes() < (size_t)n_tests * # ifdef VERY_SMALL_CONFIG 2700000 # else 33500000 # endif ) { GC_printf("Incorrect execution - missed some allocations\n"); FAIL; } if (GC_get_heap_size() + GC_get_unmapped_bytes() > max_heap_sz && !GC_get_find_leak()) { GC_printf("Unexpected heap growth - collector may be broken" " (heapsize: %lu, expected: %lu)\n", (unsigned long)(GC_get_heap_size() + GC_get_unmapped_bytes()), (unsigned long)max_heap_sz); FAIL; } # ifdef USE_MUNMAP GC_printf("Obtained %lu bytes from OS (of which %lu bytes unmapped)\n", (unsigned long)GC_get_obtained_from_os_bytes(), (unsigned long)GC_get_unmapped_bytes()); # else GC_printf("Obtained %lu bytes from OS\n", (unsigned long)GC_get_obtained_from_os_bytes()); # endif GC_printf("Final number of reachable objects is %u\n", obj_count); # ifndef GC_GET_HEAP_USAGE_NOT_NEEDED /* Get global counters (just to check the functions work). */ GC_get_heap_usage_safe(NULL, NULL, NULL, NULL, NULL); { struct GC_prof_stats_s stats; (void)GC_get_prof_stats(&stats, sizeof(stats)); # ifdef THREADS (void)GC_get_prof_stats_unsafe(&stats, sizeof(stats)); # endif } (void)GC_get_size_map_at(-1); (void)GC_get_size_map_at(1); # endif # ifndef NO_CLOCK GC_printf("Full/world-stopped collections took %lu/%lu ms\n", GC_get_full_gc_total_time(), GC_get_stopped_mark_total_time()); # endif # ifdef PARALLEL_MARK GC_printf("Completed %u collections (using %d marker threads)\n", (unsigned)GC_get_gc_no(), GC_get_parallel() + 1); # else GC_printf("Completed %u collections\n", (unsigned)GC_get_gc_no()); # endif GC_printf("Collector appears to work\n"); } #if defined(MACOS) void SetMinimumStack(long minSize) { if (minSize > LMGetDefltStack()) { long newApplLimit = (long) GetApplLimit() - (minSize - LMGetDefltStack()); SetApplLimit((Ptr) newApplLimit); MaxApplZone(); } } #define cMinStackSpace (512L * 1024L) #endif void GC_CALLBACK warn_proc(char *msg, GC_word p) { GC_printf(msg, (unsigned long)p); /*FAIL;*/ } void enable_incremental_mode(void) { # if (defined(TEST_DEFAULT_VDB) || defined(TEST_MANUAL_VDB) \ || !defined(DEFAULT_VDB)) && !defined(GC_DISABLE_INCREMENTAL) # if !defined(MAKE_BACK_GRAPH) && !defined(NO_INCREMENTAL) \ && !defined(REDIRECT_MALLOC) && !defined(USE_PROC_FOR_LIBRARIES) GC_enable_incremental(); # endif if (GC_is_incremental_mode()) { # ifndef SMALL_CONFIG if (GC_get_manual_vdb_allowed()) { GC_printf("Switched to incremental mode (manual VDB)\n"); } else # endif /* else */ { GC_printf("Switched to incremental mode\n"); if (GC_incremental_protection_needs() == GC_PROTECTS_NONE) { # if defined(PROC_VDB) || defined(SOFT_VDB) GC_printf("Reading dirty bits from /proc\n"); # elif defined(GWW_VDB) GC_printf("Using GetWriteWatch-based implementation\n"); # endif } else { GC_printf("Emulating dirty bits with mprotect/signals\n"); } } } # endif } #if defined(CPPCHECK) # define UNTESTED(sym) GC_noop1((word)&sym) #endif #if defined(MSWINCE) && defined(UNDER_CE) # define WINMAIN_LPTSTR LPWSTR #else # define WINMAIN_LPTSTR LPSTR #endif #if !defined(PCR) && !defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) #if defined(_DEBUG) && (_MSC_VER >= 1900) /* VS 2015+ */ # ifndef _CRTDBG_MAP_ALLOC # define _CRTDBG_MAP_ALLOC # endif # include /* Ensure that there is no system-malloc-allocated objects at normal */ /* exit (i.e. no such memory leaked). */ # define CRTMEM_CHECK_INIT() \ (void)_CrtSetDbgFlag(_CRTDBG_ALLOC_MEM_DF | _CRTDBG_LEAK_CHECK_DF) # define CRTMEM_DUMP_LEAKS() \ do { \ if (_CrtDumpMemoryLeaks()) { \ GC_printf("System-malloc-allocated memory leaked\n"); \ FAIL; \ } \ } while (0) #else # define CRTMEM_CHECK_INIT() (void)0 # define CRTMEM_DUMP_LEAKS() (void)0 #endif /* !_MSC_VER */ #if ((defined(MSWIN32) && !defined(__MINGW32__)) || defined(MSWINCE)) \ && !defined(NO_WINMAIN_ENTRY) int APIENTRY WinMain(HINSTANCE instance, HINSTANCE prev, WINMAIN_LPTSTR cmd, int n) #elif defined(RTEMS) # include # define CONFIGURE_APPLICATION_NEEDS_CLOCK_DRIVER # define CONFIGURE_APPLICATION_NEEDS_CONSOLE_DRIVER # define CONFIGURE_RTEMS_INIT_TASKS_TABLE # define CONFIGURE_MAXIMUM_TASKS 1 # define CONFIGURE_INIT # define CONFIGURE_INIT_TASK_STACK_SIZE (64*1024) # include rtems_task Init(rtems_task_argument ignord) #else int main(void) #endif { CRTMEM_CHECK_INIT(); # if ((defined(MSWIN32) && !defined(__MINGW32__)) || defined(MSWINCE)) \ && !defined(NO_WINMAIN_ENTRY) UNUSED_ARG(instance); UNUSED_ARG(prev); UNUSED_ARG(cmd); UNUSED_ARG(n); # if defined(CPPCHECK) GC_noop1((GC_word)&WinMain); # endif # elif defined(RTEMS) UNUSED_ARG(ignord); # if defined(CPPCHECK) GC_noop1((GC_word)&Init); # endif # endif n_tests = 0; GC_clear_exclusion_table(); /* no-op as called before GC init */ # if defined(MACOS) /* Make sure we have lots and lots of stack space. */ SetMinimumStack(cMinStackSpace); /* Cheat and let stdio initialize toolbox for us. */ printf("Testing GC Macintosh port\n"); # endif GC_COND_INIT(); GC_set_warn_proc(warn_proc); enable_incremental_mode(); set_print_procs(); GC_start_incremental_collection(); run_one_test(); # if NTHREADS > 0 { int i; for (i = 0; i < NTHREADS; i++) run_one_test(); } # endif run_single_threaded_test(); check_heap_stats(); # ifndef MSWINCE fflush(stdout); # endif # if defined(CPPCHECK) /* Entry points we should be testing, but aren't. */ # ifdef AMIGA # ifdef GC_AMIGA_FASTALLOC UNTESTED(GC_amiga_get_mem); # endif # ifndef GC_AMIGA_ONLYFAST UNTESTED(GC_amiga_set_toany); # endif # endif # if defined(MACOS) && defined(USE_TEMPORARY_MEMORY) UNTESTED(GC_MacTemporaryNewPtr); # endif UNTESTED(GC_abort_on_oom); UNTESTED(GC_deinit); # ifndef NO_DEBUGGING UNTESTED(GC_dump); UNTESTED(GC_dump_regions); UNTESTED(GC_print_free_list); # endif # endif # if !defined(GC_ANDROID_LOG) && !defined(MACOS) && !defined(OS2) \ && !defined(MSWIN32) && !defined(MSWINCE) GC_set_log_fd(2); # endif # if defined(MSWIN32) || defined(MSWINCE) || defined(CYGWIN32) GC_win32_free_heap(); # endif CRTMEM_DUMP_LEAKS(); # ifdef RTEMS exit(0); # else return 0; # endif } # endif /* !GC_WIN32_THREADS && !GC_PTHREADS */ #if defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) DWORD __stdcall thr_run_one_test(void *arg) { UNUSED_ARG(arg); run_one_test(); return 0; } #ifdef MSWINCE HANDLE win_created_h; HWND win_handle; LRESULT CALLBACK window_proc(HWND hwnd, UINT uMsg, WPARAM wParam, LPARAM lParam) { LRESULT ret = 0; switch (uMsg) { case WM_HIBERNATE: GC_printf("Received WM_HIBERNATE, calling GC_gcollect\n"); /* Force "unmap as much memory as possible" mode. */ GC_gcollect_and_unmap(); break; case WM_CLOSE: GC_printf("Received WM_CLOSE, closing window\n"); DestroyWindow(hwnd); break; case WM_DESTROY: PostQuitMessage(0); break; default: ret = DefWindowProc(hwnd, uMsg, wParam, lParam); break; } return ret; } DWORD __stdcall thr_window(void *arg) { WNDCLASS win_class = { CS_NOCLOSE, window_proc, 0, 0, GetModuleHandle(NULL), NULL, NULL, (HBRUSH)(COLOR_APPWORKSPACE+1), NULL, TEXT("GCtestWindow") }; MSG msg; UNUSED_ARG(arg); if (!RegisterClass(&win_class)) { GC_printf("RegisterClass failed\n"); FAIL; } win_handle = CreateWindowEx( 0, TEXT("GCtestWindow"), TEXT("GCtest"), 0, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, NULL, NULL, GetModuleHandle(NULL), NULL); if (NULL == win_handle) { GC_printf("CreateWindow failed\n"); FAIL; } SetEvent(win_created_h); ShowWindow(win_handle, SW_SHOW); UpdateWindow(win_handle); while (GetMessage(&msg, NULL, 0, 0)) { TranslateMessage(&msg); DispatchMessage(&msg); } return 0; } #endif #if !defined(NO_WINMAIN_ENTRY) int APIENTRY WinMain(HINSTANCE instance, HINSTANCE prev, WINMAIN_LPTSTR cmd, int n) #else int main(void) #endif { # if NTHREADS > 0 HANDLE h[NTHREADS]; int i; # endif # ifdef MSWINCE HANDLE win_thr_h; # endif DWORD thread_id; # if !defined(NO_WINMAIN_ENTRY) UNUSED_ARG(instance); UNUSED_ARG(prev); UNUSED_ARG(cmd); UNUSED_ARG(n); # if defined(CPPCHECK) GC_noop1((GC_word)&WinMain); # endif # endif # if defined(GC_DLL) && !defined(GC_NO_THREADS_DISCOVERY) \ && !defined(MSWINCE) && !defined(THREAD_LOCAL_ALLOC) GC_use_threads_discovery(); /* Test with implicit thread registration if possible. */ GC_printf("Using DllMain to track threads\n"); # endif GC_COND_INIT(); enable_incremental_mode(); InitializeCriticalSection(&incr_cs); GC_set_warn_proc(warn_proc); # ifdef MSWINCE win_created_h = CreateEvent(NULL, FALSE, FALSE, NULL); if (win_created_h == (HANDLE)NULL) { GC_printf("Event creation failed, errcode= %d\n", (int)GetLastError()); FAIL; } win_thr_h = CreateThread(NULL, 0, thr_window, 0, 0, &thread_id); if (win_thr_h == (HANDLE)NULL) { GC_printf("Thread creation failed, errcode= %d\n", (int)GetLastError()); FAIL; } if (WaitForSingleObject(win_created_h, INFINITE) != WAIT_OBJECT_0) { GC_printf("WaitForSingleObject failed 1\n"); FAIL; } CloseHandle(win_created_h); # endif set_print_procs(); # if NTHREADS > 0 for (i = 0; i < NTHREADS; i++) { h[i] = CreateThread(NULL, 0, thr_run_one_test, 0, 0, &thread_id); if (h[i] == (HANDLE)NULL) { GC_printf("Thread creation failed, errcode= %d\n", (int)GetLastError()); FAIL; } } # endif /* NTHREADS > 0 */ run_one_test(); # if NTHREADS > 0 for (i = 0; i < NTHREADS; i++) { if (WaitForSingleObject(h[i], INFINITE) != WAIT_OBJECT_0) { GC_printf("Thread wait failed, errcode= %d\n", (int)GetLastError()); FAIL; } } # endif /* NTHREADS > 0 */ # ifdef MSWINCE PostMessage(win_handle, WM_CLOSE, 0, 0); if (WaitForSingleObject(win_thr_h, INFINITE) != WAIT_OBJECT_0) { GC_printf("WaitForSingleObject failed 2\n"); FAIL; } # endif run_single_threaded_test(); check_heap_stats(); (void)GC_unregister_my_thread(); /* just to check it works (for main) */ return 0; } #endif /* GC_WIN32_THREADS */ #ifdef PCR int test(void) { PCR_Th_T * th1; PCR_Th_T * th2; int code; # if defined(CPPCHECK) GC_noop1((word)&PCR_GC_Run); GC_noop1((word)&PCR_GC_Setup); GC_noop1((word)&test); # endif n_tests = 0; /* GC_enable_incremental(); */ GC_set_warn_proc(warn_proc); set_print_procs(); th1 = PCR_Th_Fork(run_one_test, 0); th2 = PCR_Th_Fork(run_one_test, 0); run_one_test(); if (PCR_Th_T_Join(th1, &code, NIL, PCR_allSigsBlocked, PCR_waitForever) != PCR_ERes_okay || code != 0) { GC_printf("Thread 1 failed\n"); } if (PCR_Th_T_Join(th2, &code, NIL, PCR_allSigsBlocked, PCR_waitForever) != PCR_ERes_okay || code != 0) { GC_printf("Thread 2 failed\n"); } run_single_threaded_test(); check_heap_stats(); return 0; } #endif #if defined(GC_PTHREADS) # include /* for EAGAIN */ void * thr_run_one_test(void *arg) { UNUSED_ARG(arg); run_one_test(); return 0; } void GC_CALLBACK describe_norm_type(void *p, char *out_buf) { UNUSED_ARG(p); BCOPY("NORMAL", out_buf, sizeof("NORMAL")); } int GC_CALLBACK has_static_roots(const char *dlpi_name, void *section_start, size_t section_size) { UNUSED_ARG(dlpi_name); UNUSED_ARG(section_start); UNUSED_ARG(section_size); return 1; } #ifdef GC_DEBUG # define GC_free GC_debug_free #endif int main(void) { # if NTHREADS > 0 pthread_t th[NTHREADS]; int i, nthreads; # endif pthread_attr_t attr; int code; # ifdef GC_IRIX_THREADS /* Force a larger stack to be preallocated */ /* Since the initial can't always grow later. */ *((volatile char *)&code - 1024*1024) = 0; /* Require 1 MB */ # endif /* GC_IRIX_THREADS */ # if defined(GC_HPUX_THREADS) /* Default stack size is too small, especially with the 64 bit ABI */ /* Increase it. */ if (pthread_default_stacksize_np(1024*1024, 0) != 0) { GC_printf("pthread_default_stacksize_np failed\n"); } # endif /* GC_HPUX_THREADS */ # ifdef PTW32_STATIC_LIB pthread_win32_process_attach_np(); pthread_win32_thread_attach_np(); # endif # if defined(GC_DARWIN_THREADS) && !defined(GC_NO_THREADS_DISCOVERY) \ && !defined(DARWIN_DONT_PARSE_STACK) && !defined(THREAD_LOCAL_ALLOC) /* Test with the Darwin implicit thread registration. */ GC_use_threads_discovery(); GC_printf("Using Darwin task-threads-based world stop and push\n"); # endif GC_set_markers_count(0); # ifdef TEST_REUSE_SIG_SUSPEND GC_set_suspend_signal(GC_get_thr_restart_signal()); # else GC_set_suspend_signal(GC_get_suspend_signal()); # endif GC_COND_INIT(); if ((code = pthread_attr_init(&attr)) != 0) { GC_printf("pthread_attr_init failed, errno= %d\n", code); FAIL; } # if defined(GC_IRIX_THREADS) || defined(GC_FREEBSD_THREADS) \ || defined(GC_DARWIN_THREADS) || defined(GC_AIX_THREADS) \ || defined(GC_OPENBSD_THREADS) if ((code = pthread_attr_setstacksize(&attr, 1000 * 1024)) != 0) { GC_printf("pthread_attr_setstacksize failed, errno= %d\n", code); FAIL; } # endif n_tests = 0; enable_incremental_mode(); GC_set_min_bytes_allocd(1); if (GC_get_min_bytes_allocd() != 1) { GC_printf("GC_get_min_bytes_allocd() wrong result\n"); FAIL; } GC_set_rate(10); GC_set_max_prior_attempts(GC_get_max_prior_attempts()); if (GC_get_rate() != 10) { GC_printf("GC_get_rate() wrong result\n"); FAIL; } GC_set_warn_proc(warn_proc); if ((code = pthread_key_create(&fl_key, 0)) != 0) { GC_printf("Key creation failed, errno= %d\n", code); FAIL; } set_print_procs(); /* Minimal testing of some API functions. */ GC_exclude_static_roots((void *)&atomic_count, (void *)((word)&atomic_count + sizeof(atomic_count))); GC_register_has_static_roots_callback(has_static_roots); GC_register_describe_type_fn(GC_I_NORMAL, describe_norm_type); # ifdef GC_GCJ_SUPPORT (void)GC_new_proc(fake_gcj_mark_proc); # endif # if NTHREADS > 0 for (i = 0; i < NTHREADS; ++i) { if ((code = pthread_create(th+i, &attr, thr_run_one_test, 0)) != 0) { GC_printf("Thread #%d creation failed, errno= %d\n", i, code); if (i > 0 && EAGAIN == code) break; /* Resource temporarily unavailable */ FAIL; } } nthreads = i; for (; i <= NTHREADS; i++) run_one_test(); for (i = 0; i < nthreads; ++i) { if ((code = pthread_join(th[i], 0)) != 0) { GC_printf("Thread #%d join failed, errno= %d\n", i, code); FAIL; } } # else run_one_test(); # endif run_single_threaded_test(); # ifdef TRACE_BUF GC_print_trace(0); # endif check_heap_stats(); (void)fflush(stdout); (void)pthread_attr_destroy(&attr); /* Dummy checking of various getters and setters. */ (void)GC_get_bytes_since_gc(); (void)GC_get_free_bytes(); (void)GC_get_hblk_size(); (void)GC_get_is_valid_displacement_print_proc(); (void)GC_get_is_visible_print_proc(); (void)GC_get_pages_executable(); (void)GC_get_warn_proc(); (void)GC_is_disabled(); GC_set_allocd_bytes_per_finalizer(GC_get_allocd_bytes_per_finalizer()); GC_set_disable_automatic_collection(GC_get_disable_automatic_collection()); GC_set_dont_expand(GC_get_dont_expand()); GC_set_dont_precollect(GC_get_dont_precollect()); GC_set_finalize_on_demand(GC_get_finalize_on_demand()); GC_set_finalizer_notifier(GC_get_finalizer_notifier()); GC_set_force_unmap_on_gcollect(GC_get_force_unmap_on_gcollect()); GC_set_free_space_divisor(GC_get_free_space_divisor()); GC_set_full_freq(GC_get_full_freq()); GC_set_java_finalization(GC_get_java_finalization()); GC_set_max_retries(GC_get_max_retries()); GC_set_no_dls(GC_get_no_dls()); GC_set_non_gc_bytes(GC_get_non_gc_bytes()); GC_set_on_collection_event(GC_get_on_collection_event()); GC_set_on_heap_resize(GC_get_on_heap_resize()); GC_set_on_mark_stack_empty(GC_get_on_mark_stack_empty()); GC_set_on_thread_event(GC_get_on_thread_event()); GC_set_oom_fn(GC_get_oom_fn()); GC_set_push_other_roots(GC_get_push_other_roots()); GC_set_same_obj_print_proc(GC_get_same_obj_print_proc()); GC_set_start_callback(GC_get_start_callback()); GC_set_stop_func(GC_get_stop_func()); GC_set_thr_restart_signal(GC_get_thr_restart_signal()); GC_set_time_limit(GC_get_time_limit()); # if !defined(PCR) && !defined(SMALL_CONFIG) GC_set_abort_func(GC_get_abort_func()); # endif # ifndef NO_CLOCK GC_set_time_limit_tv(GC_get_time_limit_tv()); # endif # ifndef GC_NO_FINALIZATION GC_set_await_finalize_proc(GC_get_await_finalize_proc()); GC_set_interrupt_finalizers(GC_get_interrupt_finalizers()); # ifndef GC_TOGGLE_REFS_NOT_NEEDED GC_set_toggleref_func(GC_get_toggleref_func()); # endif # endif # if defined(CPPCHECK) UNTESTED(GC_register_altstack); # endif /* CPPCHECK */ # if !defined(GC_NO_DLOPEN) && !defined(DARWIN) \ && !defined(GC_WIN32_THREADS) { void *h = GC_dlopen("libc.so", 0 /* some value (maybe invalid) */); if (h != NULL) dlclose(h); } # endif # ifndef GC_NO_PTHREAD_SIGMASK { sigset_t blocked; if (GC_pthread_sigmask(SIG_BLOCK, NULL, &blocked) != 0 || GC_pthread_sigmask(SIG_BLOCK, &blocked, NULL) != 0) { GC_printf("pthread_sigmask failed\n"); FAIL; } } # endif GC_stop_world_external(); GC_start_world_external(); # if !defined(GC_NO_FINALIZATION) && !defined(JAVA_FINALIZATION_NOT_NEEDED) GC_finalize_all(); # endif GC_clear_roots(); # ifdef PTW32_STATIC_LIB pthread_win32_thread_detach_np(); pthread_win32_process_detach_np(); # else (void)GC_unregister_my_thread(); # endif return 0; } #endif /* GC_PTHREADS */