summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorBen Gamari <ben@smart-cactus.org>2021-07-12 17:26:47 -0400
committerMarge Bot <ben+marge-bot@smart-cactus.org>2021-07-27 04:47:51 -0400
commit2ce48fe9d1334dadf6fa1411238e52d65b66ba41 (patch)
tree1f48a9e2e2ae3350c3802875bef128327afdf593
parent246f08ac032392f808c7aa53dd78a96d9b43e63f (diff)
downloadhaskell-2ce48fe9d1334dadf6fa1411238e52d65b66ba41.tar.gz
rts: Break up adjustor logic
-rw-r--r--configure.ac54
-rw-r--r--distrib/configure.ac.in3
-rw-r--r--hadrian/cfg/system.config.in1
-rw-r--r--hadrian/src/Oracles/Flag.hs7
-rw-r--r--hadrian/src/Oracles/Setting.hs6
-rw-r--r--hadrian/src/Rules/Generate.hs2
-rw-r--r--hadrian/src/Settings/Packages.hs11
-rw-r--r--includes/ghc.mk2
-rw-r--r--includes/rts/storage/GC.h3
-rw-r--r--mk/config.mk.in12
-rw-r--r--rts/Adjustor.c1304
-rw-r--r--rts/Adjustor.h7
-rw-r--r--rts/adjustor/LibffiAdjustor.c176
-rw-r--r--rts/adjustor/NativeAlpha.c110
-rw-r--r--rts/adjustor/NativeAmd64.c339
-rw-r--r--rts/adjustor/NativeIA64.c165
-rw-r--r--rts/adjustor/NativePowerPC.c408
-rw-r--r--rts/adjustor/NativeSparc.c110
-rw-r--r--rts/adjustor/Nativei386.c137
-rw-r--r--rts/ghc.mk30
-rw-r--r--rts/rts.cabal.in25
21 files changed, 1583 insertions, 1329 deletions
diff --git a/configure.ac b/configure.ac
index 38de5953b6..535a0c3b88 100644
--- a/configure.ac
+++ b/configure.ac
@@ -1282,6 +1282,60 @@ case ${TargetOS} in
AC_DEFINE_UNQUOTED([RTS_LINKER_USE_MMAP], [$RtsLinkerUseMmap],
[Use mmap in the runtime linker])
+
+dnl ** Use libffi for adjustors?
+dnl --------------------------------------------------------------
+
+case ${TargetArch} in
+ i386|x86_64)
+ # We have native adjustor support on these platforms
+ HaveNativeAdjustor=yes
+ ;;
+ *)
+ HaveNativeAdjustor=no
+ ;;
+esac
+
+AC_ARG_ENABLE(libffi-adjustors,
+ [AS_HELP_STRING(
+ [--enable-libffi-adjustors],
+ [Force use of libffi for adjustors, even on platforms which have support for more efficient, native adjustors.])],
+ UseLibffiForAdjustors=$enableval,
+ dnl do nothing
+)
+
+AC_MSG_CHECKING([whether to use libffi for adjustors])
+AC_MSG_RESULT([$UseLibffiForAdjustors])
+if test "$UseLibffiForAdjustors" = "yes" ; then
+ # Use libffi is the user explicitly requested it
+ AdjustorType="libffi"
+elif test "$HaveNativeAdjustor" = "yes"; then
+ # Otherwise if we have a native adjustor implementation use that
+ AdjustorType="native"
+else
+ # If we don't have a native adjustor implementation then default to libffi
+ AdjustorType="libffi"
+fi
+
+case "$AdjustorType" in
+libffi)
+ CabalLibffiAdjustors=True
+ UseLibffiForAdjustors=YES
+ ;;
+native)
+ CabalLibffiAdjustors=False
+ UseLibffiForAdjustors=NO
+ ;;
+*)
+ AC_MSG_ERROR([Internal error: Invalid AdjustorType])
+ exit 1
+esac
+
+AC_SUBST([CabalLibffiAdjustors])
+AC_SUBST([UseLibffiForAdjustors])
+
+dnl ** Other RTS features
+dnl --------------------------------------------------------------
FP_FIND_LIBDW
FP_FIND_LIBNUMA
diff --git a/distrib/configure.ac.in b/distrib/configure.ac.in
index c287c3368d..be3a390644 100644
--- a/distrib/configure.ac.in
+++ b/distrib/configure.ac.in
@@ -31,6 +31,9 @@ LibdwLibDir=@LibdwLibDir@
AC_SUBST(LibdwLibDir)
AC_SUBST(LibdwIncludeDir)
+UseLibffiForAdjustors=@UseLibffiForAdjustors@
+AC_SUBST(UseLibffiForAdjustors)
+
# We have to run these unconditionally as FPTOOLS_SET_PLATFORM_VARS wants the
# values it computes.
AC_CANONICAL_BUILD
diff --git a/hadrian/cfg/system.config.in b/hadrian/cfg/system.config.in
index 4d9d1b752e..e23c79a81d 100644
--- a/hadrian/cfg/system.config.in
+++ b/hadrian/cfg/system.config.in
@@ -47,6 +47,7 @@ hs-cpp-args = @HaskellCPPArgs@
solaris-broken-shld = @SOLARIS_BROKEN_SHLD@
ghc-unregisterised = @Unregisterised@
tables-next-to-code = @TablesNextToCode@
+use-libffi-for-adjustors = @UseLibffiForAdjustors@
ghc-source-path = @hardtop@
leading-underscore = @LeadingUnderscore@
diff --git a/hadrian/src/Oracles/Flag.hs b/hadrian/src/Oracles/Flag.hs
index 1b3aff693e..5b41698be1 100644
--- a/hadrian/src/Oracles/Flag.hs
+++ b/hadrian/src/Oracles/Flag.hs
@@ -2,7 +2,7 @@
module Oracles.Flag (
Flag (..), flag, getFlag, platformSupportsSharedLibs,
- targetSupportsSMP
+ targetSupportsSMP, useLibffiForAdjustors
) where
import Hadrian.Oracles.TextFile
@@ -26,6 +26,7 @@ data Flag = ArSupportsAtFile
| UseSystemFfi
| BootstrapThreadedRts
| SystemDistroMINGW
+ | UseLibffiForAdjustors
-- Note, if a flag is set to empty string we treat it as set to NO. This seems
-- fragile, but some flags do behave like this.
@@ -47,6 +48,7 @@ flag f = do
UseSystemFfi -> "use-system-ffi"
BootstrapThreadedRts -> "bootstrap-threaded-rts"
SystemDistroMINGW -> "system-use-distro-mingw"
+ UseLibffiForAdjustors -> "use-libffi-for-adjustors"
value <- lookupValueOrError configFile key
when (value `notElem` ["YES", "NO", ""]) . error $ "Configuration flag "
++ quote (key ++ " = " ++ value) ++ " cannot be parsed."
@@ -88,3 +90,6 @@ targetSupportsSMP = do
, ver < ARMv7 -> return False
| goodArch -> return True
| otherwise -> return False
+
+useLibffiForAdjustors :: Action Bool
+useLibffiForAdjustors = flag UseLibffiForAdjustors
diff --git a/hadrian/src/Oracles/Setting.hs b/hadrian/src/Oracles/Setting.hs
index 3713b2457a..73fd1ce273 100644
--- a/hadrian/src/Oracles/Setting.hs
+++ b/hadrian/src/Oracles/Setting.hs
@@ -14,7 +14,7 @@ module Oracles.Setting (
isElfTarget,
ArmVersion(..),
targetArmVersion,
- ghcWithInterpreter, useLibFFIForAdjustors
+ ghcWithInterpreter
) where
import Hadrian.Expression
@@ -275,10 +275,6 @@ ghcWithInterpreter = do
, "powerpc64", "powerpc64le" ]
return $ goodOs && goodArch
--- | Check to use @libffi@ for adjustors.
-useLibFFIForAdjustors :: Action Bool
-useLibFFIForAdjustors = notM $ anyTargetArch ["i386", "x86_64"]
-
-- | Variants of the ARM architecture.
data ArmVersion = ARMv5 | ARMv6 | ARMv7
deriving (Eq, Ord, Show, Read)
diff --git a/hadrian/src/Rules/Generate.hs b/hadrian/src/Rules/Generate.hs
index b7b4d17180..3eb100bae5 100644
--- a/hadrian/src/Rules/Generate.hs
+++ b/hadrian/src/Rules/Generate.hs
@@ -341,7 +341,7 @@ generateSettings = do
, ("RTS ways", unwords . map show <$> getRtsWays)
, ("Tables next to code", expr $ yesNo <$> flag TablesNextToCode)
, ("Leading underscore", expr $ yesNo <$> flag LeadingUnderscore)
- , ("Use LibFFI", expr $ yesNo <$> useLibFFIForAdjustors)
+ , ("Use LibFFI", expr $ yesNo <$> useLibffiForAdjustors)
, ("RTS expects libdw", yesNo <$> getFlag WithLibdw)
]
let showTuple (k, v) = "(" ++ show k ++ ", " ++ show v ++ ")"
diff --git a/hadrian/src/Settings/Packages.hs b/hadrian/src/Settings/Packages.hs
index 8161598431..9bf82a6cef 100644
--- a/hadrian/src/Settings/Packages.hs
+++ b/hadrian/src/Settings/Packages.hs
@@ -294,8 +294,6 @@ rtsPackageArgs = package rts ? do
, "-g3"
, "-O0" ]
- , useLibFFIForAdjustors ? arg "-DUSE_LIBFFI_FOR_ADJUSTORS"
-
, inputs ["**/RtsMessages.c", "**/Trace.c"] ?
arg ("-DProjectVersion=" ++ show projectVersion)
@@ -356,10 +354,11 @@ rtsPackageArgs = package rts ? do
mconcat
[ builder (Cabal Flags) ? mconcat
[ any (wayUnit Profiling) rtsWays ? arg "profiling"
- , any (wayUnit Debug) rtsWays ? arg "debug"
- , any (wayUnit Logging) rtsWays ? arg "logging"
- , any (wayUnit Dynamic) rtsWays ? arg "dynamic"
- , Debug `wayUnit` way ? arg "find-ptr"
+ , any (wayUnit Debug) rtsWays ? arg "debug"
+ , any (wayUnit Logging) rtsWays ? arg "logging"
+ , any (wayUnit Dynamic) rtsWays ? arg "dynamic"
+ , useLibffiForAdjustors ? arg "libffi-adjustors"
+ , Debug `wayUnit` way ? arg "find-ptr"
]
, builder (Cabal Setup) ? mconcat
[ if not (null libdwLibraryDir) then arg ("--extra-lib-dirs="++libdwLibraryDir) else mempty
diff --git a/includes/ghc.mk b/includes/ghc.mk
index 00e518c34d..5e3060c238 100644
--- a/includes/ghc.mk
+++ b/includes/ghc.mk
@@ -258,7 +258,7 @@ $(includes_SETTINGS) : includes/Makefile | $$(dir $$@)/.
@echo ',("RTS ways", "$(GhcRTSWays)")' >> $@
@echo ',("Tables next to code", "$(TablesNextToCode)")' >> $@
@echo ',("Leading underscore", "$(LeadingUnderscore)")' >> $@
- @echo ',("Use LibFFI", "$(UseLibFFIForAdjustors)")' >> $@
+ @echo ',("Use LibFFI", "$(UseLibffiForAdjustors)")' >> $@
@echo ",(\"RTS expects libdw\", \"$(GhcRtsWithLibdw)\")" >> $@
@echo "]" >> $@
diff --git a/includes/rts/storage/GC.h b/includes/rts/storage/GC.h
index 0258811728..6f8f382d56 100644
--- a/includes/rts/storage/GC.h
+++ b/includes/rts/storage/GC.h
@@ -205,9 +205,6 @@ typedef void* AdjustorExecutable;
AdjustorWritable allocateExec(W_ len, AdjustorExecutable *exec_addr);
void flushExec(W_ len, AdjustorExecutable exec_addr);
-#if defined(darwin_HOST_OS)
-AdjustorWritable execToWritable(AdjustorExecutable exec);
-#endif
#if RTS_LINKER_USE_MMAP
AdjustorWritable allocateWrite(W_ bytes);
void markExec(W_ bytes, AdjustorWritable writ);
diff --git a/mk/config.mk.in b/mk/config.mk.in
index 086a9147f2..35f6e2d087 100644
--- a/mk/config.mk.in
+++ b/mk/config.mk.in
@@ -181,16 +181,6 @@ else
GhcWithInterpreter=$(if $(findstring YES,$(DYNAMIC_GHC_PROGRAMS)),YES,NO)
endif
-# Whether to use libffi for adjustors (foreign import "wrapper") or
-# not. If we have built-in support (rts/Adjustor.c) then we use that,
-# otherwise we fall back on libffi, which is slightly slower.
-ArchHasAdjustorSupport = $(if $(findstring $(TargetArch_CPP),i386 x86_64),YES,NO)
-ifeq "$(ArchHasAdjustorSupport)" "YES"
-UseLibFFIForAdjustors=NO
-else
-UseLibFFIForAdjustors=YES
-endif
-
# On Windows we normally want to make a relocatable bindist, to we
# ignore flags like libdir
ifeq "$(Windows_Host)" "YES"
@@ -333,6 +323,8 @@ UseSystemLibFFI=@UseSystemLibFFI@
FFILibDir=@FFILibDir@
FFIIncludeDir=@FFIIncludeDir@
+UseLibffiForAdjustors=@UseLibffiForAdjustors@
+
UseLibdw=@UseLibdw@
LibdwLibDir=@LibdwLibDir@
LibdwIncludeDir=@LibdwIncludeDir@
diff --git a/rts/Adjustor.c b/rts/Adjustor.c
index ecd37f2d10..974d8e3d37 100644
--- a/rts/Adjustor.c
+++ b/rts/Adjustor.c
@@ -36,332 +36,9 @@ and C resources. Failure to do so will result in memory leaks on both the C and
Haskell side.
*/
-#include "PosixSource.h"
-#include "Rts.h"
+#include "Adjustor.h"
-#include "RtsUtils.h"
-#include "StablePtr.h"
-
-#if defined(USE_LIBFFI_FOR_ADJUSTORS)
-#include "ffi.h"
-#include <string.h>
-#endif
-
-#if defined(i386_HOST_ARCH)
-extern void adjustorCode(void);
-#elif defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH)
-// from AdjustorAsm.s
-// not declared as a function so that AIX-style
-// fundescs can never get in the way.
-extern void *adjustorCode;
-#endif
-
-#if defined(USE_LIBFFI_FOR_ADJUSTORS)
-
-/* Maps AdjustorExecutable* to AdjustorWritable*. */
-static HashTable* allocatedExecs;
-
-static AdjustorWritable allocate_adjustor(AdjustorExecutable *exec_ret)
-{
- AdjustorWritable writ;
- ffi_closure* cl;
-
- ACQUIRE_SM_LOCK;
- cl = writ = ffi_closure_alloc(sizeof(ffi_closure), exec_ret);
- if (cl != NULL) {
- if (allocatedExecs == NULL) {
- allocatedExecs = allocHashTable();
- }
- insertHashTable(allocatedExecs, (StgWord)*exec_ret, writ);
- }
- RELEASE_SM_LOCK;
- return writ;
-}
-
-static AdjustorWritable exec_to_writable(AdjustorExecutable exec)
-{
- AdjustorWritable writ;
- ACQUIRE_SM_LOCK;
- if (allocatedExecs == NULL ||
- (writ = lookupHashTable(allocatedExecs, (StgWord)exec)) == NULL) {
- RELEASE_SM_LOCK;
- barf("execToWritable: not found");
- }
- RELEASE_SM_LOCK;
- return writ;
-}
-
-static void free_adjustor(AdjustorExecutable exec)
-{
- AdjustorWritable writ;
- ffi_closure* cl;
- cl = writ = execToWritable(exec);
- ACQUIRE_SM_LOCK;
- removeHashTable(allocatedExecs, (StgWord)exec, writ);
- ffi_closure_free(cl);
- RELEASE_SM_LOCK;
-}
-
-
-/* There are subtle differences between how libffi adjustors work on
- * different platforms, and the situation is a little complex.
- *
- * HOW ADJUSTORS/CLOSURES WORK ON LIBFFI:
- * libffi's ffi_closure_alloc() function gives you two pointers to a closure,
- * 1. the writable pointer, and 2. the executable pointer. You write the
- * closure into the writable pointer (and ffi_prep_closure_loc() will do this
- * for you) and you execute it at the executable pointer.
- *
- * THE PROBLEM:
- * The RTS deals only with the executable pointer, but when it comes time to
- * free the closure, libffi wants the writable pointer back that it gave you
- * when you allocated it.
- *
- * On Linux we solve this problem by storing the address of the writable
- * mapping into itself, then returning both writable and executable pointers
- * plus 1 machine word for preparing the closure for use by the RTS (see the
- * Linux version of allocateExec() in rts/sm/Storage.c). When we want to
- * recover the writable address, we subtract 1 word from the executable
- * address and fetch. This works because Linux kernel magic gives us two
- * pointers with different addresses that refer to the same memory. Whatever
- * you write into the writable address can be read back at the executable
- * address. This method is very efficient.
- *
- * On iOS this breaks for two reasons: 1. the two pointers do not refer to
- * the same memory (so we can't retrieve anything stored into the writable
- * pointer if we only have the exec pointer), and 2. libffi's
- * ffi_closure_alloc() assumes the pointer it has returned you is a
- * ffi_closure structure and treats it as such: It uses that memory to
- * communicate with ffi_prep_closure_loc(). On Linux by contrast
- * ffi_closure_alloc() is viewed simply as a memory allocation, and only
- * ffi_prep_closure_loc() deals in ffi_closure structures. Each of these
- * differences is enough make the efficient way used on Linux not work on iOS.
- * Instead on iOS we use hash tables to recover the writable address from the
- * executable one. This method is conservative and would almost certainly work
- * on any platform, but on Linux it makes sense to use the faster method.
- */
-void
-freeHaskellFunctionPtr(void* ptr)
-{
- ffi_closure *cl;
-
- cl = exec_to_writable(ptr);
- freeStablePtr(cl->user_data);
- stgFree(cl->cif->arg_types);
- stgFree(cl->cif);
- free_adjustor(ptr);
-}
-
-static ffi_type * char_to_ffi_type(char c)
-{
- switch (c) {
- case 'v': return &ffi_type_void;
- case 'f': return &ffi_type_float;
- case 'd': return &ffi_type_double;
- case 'L': return &ffi_type_sint64;
- case 'l': return &ffi_type_uint64;
- case 'W': return &ffi_type_sint32;
- case 'w': return &ffi_type_uint32;
- case 'S': return &ffi_type_sint16;
- case 's': return &ffi_type_uint16;
- case 'B': return &ffi_type_sint8;
- case 'b': return &ffi_type_uint8;
- case 'p': return &ffi_type_pointer;
- default: barf("char_to_ffi_type: unknown type '%c'", c);
- }
-}
-
-void*
-createAdjustor (int cconv,
- StgStablePtr hptr,
- StgFunPtr wptr,
- char *typeString)
-{
- ffi_cif *cif;
- ffi_type **arg_types;
- uint32_t n_args, i;
- ffi_type *result_type;
- ffi_closure *cl;
- int r, abi;
- void *code;
-
- n_args = strlen(typeString) - 1;
- cif = stgMallocBytes(sizeof(ffi_cif), "createAdjustor");
- arg_types = stgMallocBytes(n_args * sizeof(ffi_type*), "createAdjustor");
-
- result_type = char_to_ffi_type(typeString[0]);
- for (i=0; i < n_args; i++) {
- arg_types[i] = char_to_ffi_type(typeString[i+1]);
- }
- switch (cconv) {
-#if defined(mingw32_HOST_OS) && defined(i386_HOST_ARCH)
- case 0: /* stdcall */
- abi = FFI_STDCALL;
- break;
-#endif /* defined(mingw32_HOST_OS) && defined(i386_HOST_ARCH) */
- case 1: /* ccall */
- abi = FFI_DEFAULT_ABI;
- break;
- default:
- barf("createAdjustor: convention %d not supported on this platform", cconv);
- }
-
- r = ffi_prep_cif(cif, abi, n_args, result_type, arg_types);
- if (r != FFI_OK) barf("ffi_prep_cif failed: %d", r);
-
- cl = allocate_adjustor(&code);
- if (cl == NULL) {
- barf("createAdjustor: failed to allocate memory");
- }
-
- r = ffi_prep_closure_loc(cl, cif, (void*)wptr, hptr/*userdata*/, code);
- if (r != FFI_OK) barf("ffi_prep_closure_loc failed: %d", r);
-
- return (void*)code;
-}
-
-#else // To end of file...
-
-#if defined(_WIN32)
-#include <windows.h>
-#endif
-
-#if defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS)
-#include <string.h>
-#endif
-
-#if defined(LEADING_UNDERSCORE)
-#define UNDERSCORE "_"
-#else
-#define UNDERSCORE ""
-#endif
-
-#if defined(x86_64_HOST_ARCH)
-/*
- Now here's something obscure for you:
-
- When generating an adjustor thunk that uses the C calling
- convention, we have to make sure that the thunk kicks off
- the process of jumping into Haskell with a tail jump. Why?
- Because as a result of jumping in into Haskell we may end
- up freeing the very adjustor thunk we came from using
- freeHaskellFunctionPtr(). Hence, we better not return to
- the adjustor code on our way out, since it could by then
- point to junk.
-
- The fix is readily at hand, just include the opcodes
- for the C stack fixup code that we need to perform when
- returning in some static piece of memory and arrange
- to return to it before tail jumping from the adjustor thunk.
-*/
-static void GNUC3_ATTRIBUTE(used) obscure_ccall_wrapper(void)
-{
- __asm__ (
- ".globl " UNDERSCORE "obscure_ccall_ret_code\n"
- UNDERSCORE "obscure_ccall_ret_code:\n\t"
- "addq $0x8, %rsp\n\t"
-#if defined(mingw32_HOST_OS)
- /* On Win64, we had to put the original return address after the
- arg 1-4 spill slots, ro now we have to move it back */
- "movq 0x20(%rsp), %rcx\n"
- "movq %rcx, (%rsp)\n"
-#endif /* defined(mingw32_HOST_OS) */
- "ret"
- );
-}
-extern void obscure_ccall_ret_code(void);
-#endif /* defined(x86_64_HOST_ARCH) */
-
-#if defined(alpha_HOST_ARCH)
-/* To get the definition of PAL_imb: */
-# if defined(linux_HOST_OS)
-# include <asm/pal.h>
-# else
-# include <machine/pal.h>
-# endif
-#endif
-
-#if defined(ia64_HOST_ARCH)
-
-/* Layout of a function descriptor */
-typedef struct _IA64FunDesc {
- StgWord64 ip;
- StgWord64 gp;
-} IA64FunDesc;
-
-static void *
-stgAllocStable(size_t size_in_bytes, StgStablePtr *stable)
-{
- StgArrBytes* arr;
- uint32_t data_size_in_words, total_size_in_words;
-
- /* round up to a whole number of words */
- data_size_in_words = ROUNDUP_BYTES_TO_WDS(size_in_bytes);
- total_size_in_words = sizeofW(StgArrBytes) + data_size_in_words;
-
- /* allocate and fill it in */
- arr = (StgArrBytes *)allocate(total_size_in_words);
- SET_ARR_HDR(arr, &stg_ARR_WORDS_info, CCCS, size_in_bytes);
-
- /* obtain a stable ptr */
- *stable = getStablePtr((StgPtr)arr);
-
- /* and return a ptr to the goods inside the array */
- return(&(arr->payload));
-}
-#endif /* defined(ia64_HOST_ARCH) */
-
-#if defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS)
-__asm__("obscure_ccall_ret_code:\n\t"
- "lwz 1,0(1)\n\t"
- "lwz 0,4(1)\n\t"
- "mtlr 0\n\t"
- "blr");
-extern void obscure_ccall_ret_code(void);
-#endif /* defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS) */
-
-#if defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH)
-#if !(defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS))
-
-/* !!! !!! WARNING: !!! !!!
- * This structure is accessed from AdjustorAsm.s
- * Any changes here have to be mirrored in the offsets there.
- */
-
-typedef struct AdjustorStub {
- /* fundesc-based ABIs */
-#define FUNDESCS
- StgFunPtr code;
- struct AdjustorStub
- *toc;
- void *env;
- StgStablePtr hptr;
- StgFunPtr wptr;
- StgInt negative_framesize;
- StgInt extrawords_plus_one;
-} AdjustorStub;
-
-#endif /* !(defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS)) */
-#endif /* defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH) */
-
-#if defined(i386_HOST_ARCH)
-
-/* !!! !!! WARNING: !!! !!!
- * This structure is accessed from AdjustorAsm.s
- * Any changes here have to be mirrored in the offsets there.
- */
-
-typedef struct AdjustorStub {
- unsigned char call[8];
- StgStablePtr hptr;
- StgFunPtr wptr;
- StgInt frame_size;
- StgInt argument_size;
-} AdjustorStub;
-#endif /* defined(i386_HOST_ARCH) */
-
-#if defined(i386_HOST_ARCH) || defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH)
-static int totalArgumentSize(char *typeString)
+int totalArgumentSize(const char *typeString)
{
int sz = 0;
while(*typeString)
@@ -370,973 +47,20 @@ static int totalArgumentSize(char *typeString)
switch(t)
{
- // on 32-bit platforms, Double and Int64 occupy two words.
- case 'd':
- case 'l':
- case 'L':
- if(sizeof(void*) == 4)
- {
- sz += 2;
- break;
- }
- /* fall through */
- // everything else is one word.
- default:
- sz += 1;
- }
- }
- return sz;
-}
-#endif /* defined(i386_HOST_ARCH) || defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH) */
-
-void*
-createAdjustor(int cconv, StgStablePtr hptr,
- StgFunPtr wptr,
- char *typeString
-#if !defined(powerpc_HOST_ARCH) && !defined(powerpc64_HOST_ARCH) && !defined(x86_64_HOST_ARCH)
- STG_UNUSED
-#endif
- )
-{
- void *adjustor = NULL;
- void *code = NULL;
-
- switch (cconv)
- {
- case 0: /* _stdcall */
-#if defined(i386_HOST_ARCH) && !defined(darwin_HOST_OS)
- /* Magic constant computed by inspecting the code length of
- the following assembly language snippet
- (offset and machine code prefixed):
-
- <0>: 58 popl %eax # temp. remove ret addr..
- <1>: 68 fd fc fe fa pushl 0xfafefcfd # constant is large enough to
- # hold a StgStablePtr
- <6>: 50 pushl %eax # put back ret. addr
- <7>: b8 fa ef ff 00 movl $0x00ffeffa, %eax # load up wptr
- <c>: ff e0 jmp %eax # and jump to it.
- # the callee cleans up the stack
- */
- adjustor = allocateExec(14,&code);
- {
- unsigned char *const adj_code = (unsigned char *)adjustor;
- adj_code[0x00] = (unsigned char)0x58; /* popl %eax */
-
- adj_code[0x01] = (unsigned char)0x68; /* pushl hptr (which is a dword immediate ) */
- *((StgStablePtr*)(adj_code + 0x02)) = (StgStablePtr)hptr;
-
- adj_code[0x06] = (unsigned char)0x50; /* pushl %eax */
-
- adj_code[0x07] = (unsigned char)0xb8; /* movl $wptr, %eax */
- *((StgFunPtr*)(adj_code + 0x08)) = (StgFunPtr)wptr;
-
- adj_code[0x0c] = (unsigned char)0xff; /* jmp %eax */
- adj_code[0x0d] = (unsigned char)0xe0;
- }
-#endif /* defined(i386_HOST_ARCH) && !defined(darwin_HOST_OS) */
- break;
-
- case 1: /* _ccall */
-#if defined(i386_HOST_ARCH)
- {
- /*
- Most of the trickiness here is due to the need to keep the
- stack pointer 16-byte aligned (see #5250). That means we
- can't just push another argument on the stack and call the
- wrapper, we may have to shuffle the whole argument block.
-
- We offload most of the work to AdjustorAsm.S.
- */
- AdjustorStub *adjustorStub = allocateExec(sizeof(AdjustorStub),&code);
- adjustor = adjustorStub;
-
- int sz = totalArgumentSize(typeString);
-
- adjustorStub->call[0] = 0xe8;
- *(long*)&adjustorStub->call[1] = ((char*)&adjustorCode) - ((char*)code + 5);
- adjustorStub->hptr = hptr;
- adjustorStub->wptr = wptr;
-
- // The adjustor puts the following things on the stack:
- // 1.) %ebp link
- // 2.) padding and (a copy of) the arguments
- // 3.) a dummy argument
- // 4.) hptr
- // 5.) return address (for returning to the adjustor)
- // All these have to add up to a multiple of 16.
-
- // first, include everything in frame_size
- adjustorStub->frame_size = sz * 4 + 16;
- // align to 16 bytes
- adjustorStub->frame_size = (adjustorStub->frame_size + 15) & ~15;
- // only count 2.) and 3.) as part of frame_size
- adjustorStub->frame_size -= 12;
- adjustorStub->argument_size = sz;
- }
-
-#elif defined(x86_64_HOST_ARCH)
-
-# if defined(mingw32_HOST_OS)
- /*
- stack at call:
- argn
- ...
- arg5
- return address
- %rcx,%rdx,%r8,%r9 = arg1..arg4
-
- if there are <4 integer args, then we can just push the
- StablePtr into %rcx and shuffle the other args up.
-
- If there are >=4 integer args, then we have to flush one arg
- to the stack, and arrange to adjust the stack ptr on return.
- The stack will be rearranged to this:
-
- argn
- ...
- arg5
- return address *** <-- dummy arg in stub fn.
- arg4
- obscure_ccall_ret_code
-
- This unfortunately means that the type of the stub function
- must have a dummy argument for the original return address
- pointer inserted just after the 4th integer argument.
-
- Code for the simple case:
-
- 0: 4d 89 c1 mov %r8,%r9
- 3: 49 89 d0 mov %rdx,%r8
- 6: 48 89 ca mov %rcx,%rdx
- 9: f2 0f 10 da movsd %xmm2,%xmm3
- d: f2 0f 10 d1 movsd %xmm1,%xmm2
- 11: f2 0f 10 c8 movsd %xmm0,%xmm1
- 15: 48 8b 0d 0c 00 00 00 mov 0xc(%rip),%rcx # 28 <.text+0x28>
- 1c: ff 25 0e 00 00 00 jmpq *0xe(%rip) # 30 <.text+0x30>
- 22: 90 nop
- [...]
-
-
- And the version for >=4 integer arguments:
-
-[we want to push the 4th argument (either %r9 or %xmm3, depending on
- whether it is a floating arg or not) and the return address onto the
- stack. However, slots 1-4 are reserved for code we call to spill its
- args 1-4 into, so we can't just push them onto the bottom of the stack.
- So first put the 4th argument onto the stack, above what will be the
- spill slots.]
- 0: 48 83 ec 08 sub $0x8,%rsp
-[if non-floating arg, then do this:]
- 4: 90 nop
- 5: 4c 89 4c 24 20 mov %r9,0x20(%rsp)
-[else if floating arg then do this:]
- 4: f2 0f 11 5c 24 20 movsd %xmm3,0x20(%rsp)
-[end if]
-[Now push the new return address onto the stack]
- a: ff 35 30 00 00 00 pushq 0x30(%rip) # 40 <.text+0x40>
-[But the old return address has been moved up into a spill slot, so
- we need to move it above them]
- 10: 4c 8b 4c 24 10 mov 0x10(%rsp),%r9
- 15: 4c 89 4c 24 30 mov %r9,0x30(%rsp)
-[Now we do the normal register shuffle-up etc]
- 1a: 4d 89 c1 mov %r8,%r9
- 1d: 49 89 d0 mov %rdx,%r8
- 20: 48 89 ca mov %rcx,%rdx
- 23: f2 0f 10 da movsd %xmm2,%xmm3
- 27: f2 0f 10 d1 movsd %xmm1,%xmm2
- 2b: f2 0f 10 c8 movsd %xmm0,%xmm1
- 2f: 48 8b 0d 12 00 00 00 mov 0x12(%rip),%rcx # 48 <.text+0x48>
- 36: ff 25 14 00 00 00 jmpq *0x14(%rip) # 50 <.text+0x50>
- 3c: 90 nop
- 3d: 90 nop
- 3e: 90 nop
- 3f: 90 nop
- [...]
-
- */
- {
- StgWord8 *adj_code;
-
- // determine whether we have 4 or more integer arguments,
- // and therefore need to flush one to the stack.
- if ((typeString[0] == '\0') ||
- (typeString[1] == '\0') ||
- (typeString[2] == '\0') ||
- (typeString[3] == '\0')) {
-
- adjustor = allocateExec(0x38,&code);
- adj_code = (StgWord8*)adjustor;
-
- *(StgInt32 *)adj_code = 0x49c1894d;
- *(StgInt32 *)(adj_code+0x4) = 0x8948d089;
- *(StgInt32 *)(adj_code+0x8) = 0x100ff2ca;
- *(StgInt32 *)(adj_code+0xc) = 0x100ff2da;
- *(StgInt32 *)(adj_code+0x10) = 0x100ff2d1;
- *(StgInt32 *)(adj_code+0x14) = 0x0d8b48c8;
- *(StgInt32 *)(adj_code+0x18) = 0x0000000c;
-
- *(StgInt32 *)(adj_code+0x1c) = 0x000e25ff;
- *(StgInt32 *)(adj_code+0x20) = 0x00000000;
- *(StgInt64 *)(adj_code+0x28) = (StgInt64)hptr;
- *(StgInt64 *)(adj_code+0x30) = (StgInt64)wptr;
- }
- else
- {
- int fourthFloating;
-
- fourthFloating = (typeString[3] == 'f' || typeString[3] == 'd');
- adjustor = allocateExec(0x58,&code);
- adj_code = (StgWord8*)adjustor;
- *(StgInt32 *)adj_code = 0x08ec8348;
- *(StgInt32 *)(adj_code+0x4) = fourthFloating ? 0x5c110ff2
- : 0x4c894c90;
- *(StgInt32 *)(adj_code+0x8) = 0x35ff2024;
- *(StgInt32 *)(adj_code+0xc) = 0x00000030;
- *(StgInt32 *)(adj_code+0x10) = 0x244c8b4c;
- *(StgInt32 *)(adj_code+0x14) = 0x4c894c10;
- *(StgInt32 *)(adj_code+0x18) = 0x894d3024;
- *(StgInt32 *)(adj_code+0x1c) = 0xd08949c1;
- *(StgInt32 *)(adj_code+0x20) = 0xf2ca8948;
- *(StgInt32 *)(adj_code+0x24) = 0xf2da100f;
- *(StgInt32 *)(adj_code+0x28) = 0xf2d1100f;
- *(StgInt32 *)(adj_code+0x2c) = 0x48c8100f;
- *(StgInt32 *)(adj_code+0x30) = 0x00120d8b;
- *(StgInt32 *)(adj_code+0x34) = 0x25ff0000;
- *(StgInt32 *)(adj_code+0x38) = 0x00000014;
- *(StgInt32 *)(adj_code+0x3c) = 0x90909090;
- *(StgInt64 *)(adj_code+0x40) = (StgInt64)obscure_ccall_ret_code;
- *(StgInt64 *)(adj_code+0x48) = (StgInt64)hptr;
- *(StgInt64 *)(adj_code+0x50) = (StgInt64)wptr;
- }
- }
-
-# else
- /*
- stack at call:
- argn
- ...
- arg7
- return address
- %rdi,%rsi,%rdx,%rcx,%r8,%r9 = arg1..arg6
-
- if there are <6 integer args, then we can just push the
- StablePtr into %edi and shuffle the other args up.
-
- If there are >=6 integer args, then we have to flush one arg
- to the stack, and arrange to adjust the stack ptr on return.
- The stack will be rearranged to this:
-
- argn
- ...
- arg7
- return address *** <-- dummy arg in stub fn.
- arg6
- obscure_ccall_ret_code
-
- This unfortunately means that the type of the stub function
- must have a dummy argument for the original return address
- pointer inserted just after the 6th integer argument.
-
- Code for the simple case:
-
- 0: 4d 89 c1 mov %r8,%r9
- 3: 49 89 c8 mov %rcx,%r8
- 6: 48 89 d1 mov %rdx,%rcx
- 9: 48 89 f2 mov %rsi,%rdx
- c: 48 89 fe mov %rdi,%rsi
- f: 48 8b 3d 0a 00 00 00 mov 10(%rip),%rdi
- 16: ff 25 0c 00 00 00 jmpq *12(%rip)
- ...
- 20: .quad 0 # aligned on 8-byte boundary
- 28: .quad 0 # aligned on 8-byte boundary
-
-
- And the version for >=6 integer arguments:
-
- 0: 41 51 push %r9
- 2: ff 35 20 00 00 00 pushq 32(%rip) # 28 <ccall_adjustor+0x28>
- 8: 4d 89 c1 mov %r8,%r9
- b: 49 89 c8 mov %rcx,%r8
- e: 48 89 d1 mov %rdx,%rcx
- 11: 48 89 f2 mov %rsi,%rdx
- 14: 48 89 fe mov %rdi,%rsi
- 17: 48 8b 3d 12 00 00 00 mov 18(%rip),%rdi # 30 <ccall_adjustor+0x30>
- 1e: ff 25 14 00 00 00 jmpq *20(%rip) # 38 <ccall_adjustor+0x38>
- ...
- 28: .quad 0 # aligned on 8-byte boundary
- 30: .quad 0 # aligned on 8-byte boundary
- 38: .quad 0 # aligned on 8-byte boundary
- */
-
- {
- int i = 0;
- char *c;
- StgWord8 *adj_code;
-
- // determine whether we have 6 or more integer arguments,
- // and therefore need to flush one to the stack.
- for (c = typeString; *c != '\0'; c++) {
- if (*c != 'f' && *c != 'd') i++;
- if (i == 6) break;
- }
-
- if (i < 6) {
- adjustor = allocateExec(0x30,&code);
- adj_code = (StgWord8*)adjustor;
-
- *(StgInt32 *)adj_code = 0x49c1894d;
- *(StgInt32 *)(adj_code+0x4) = 0x8948c889;
- *(StgInt32 *)(adj_code+0x8) = 0xf28948d1;
- *(StgInt32 *)(adj_code+0xc) = 0x48fe8948;
- *(StgInt32 *)(adj_code+0x10) = 0x000a3d8b;
- *(StgInt32 *)(adj_code+0x14) = 0x25ff0000;
- *(StgInt32 *)(adj_code+0x18) = 0x0000000c;
- *(StgInt64 *)(adj_code+0x20) = (StgInt64)hptr;
- *(StgInt64 *)(adj_code+0x28) = (StgInt64)wptr;
- }
- else
- {
- adjustor = allocateExec(0x40,&code);
- adj_code = (StgWord8*)adjustor;
-
- *(StgInt32 *)adj_code = 0x35ff5141;
- *(StgInt32 *)(adj_code+0x4) = 0x00000020;
- *(StgInt32 *)(adj_code+0x8) = 0x49c1894d;
- *(StgInt32 *)(adj_code+0xc) = 0x8948c889;
- *(StgInt32 *)(adj_code+0x10) = 0xf28948d1;
- *(StgInt32 *)(adj_code+0x14) = 0x48fe8948;
- *(StgInt32 *)(adj_code+0x18) = 0x00123d8b;
- *(StgInt32 *)(adj_code+0x1c) = 0x25ff0000;
- *(StgInt32 *)(adj_code+0x20) = 0x00000014;
-
- *(StgInt64 *)(adj_code+0x28) = (StgInt64)obscure_ccall_ret_code;
- *(StgInt64 *)(adj_code+0x30) = (StgInt64)hptr;
- *(StgInt64 *)(adj_code+0x38) = (StgInt64)wptr;
- }
- }
-# endif
-
-
-#elif defined(sparc_HOST_ARCH)
- /* Magic constant computed by inspecting the code length of the following
- assembly language snippet (offset and machine code prefixed):
-
- <00>: 9C23A008 sub %sp, 8, %sp ! make room for %o4/%o5 in caller's frame
- <04>: DA23A060 st %o5, [%sp + 96] ! shift registers by 2 positions
- <08>: D823A05C st %o4, [%sp + 92]
- <0C>: 9A10000B mov %o3, %o5
- <10>: 9810000A mov %o2, %o4
- <14>: 96100009 mov %o1, %o3
- <18>: 94100008 mov %o0, %o2
- <1C>: 13000000 sethi %hi(wptr), %o1 ! load up wptr (1 of 2)
- <20>: 11000000 sethi %hi(hptr), %o0 ! load up hptr (1 of 2)
- <24>: 81C26000 jmp %o1 + %lo(wptr) ! jump to wptr (load 2 of 2)
- <28>: 90122000 or %o0, %lo(hptr), %o0 ! load up hptr (2 of 2, delay slot)
- <2C> 00000000 ! place for getting hptr back easily
-
- ccall'ing on SPARC is easy, because we are quite lucky to push a
- multiple of 8 bytes (1 word hptr + 1 word dummy arg) in front of the
- existing arguments (note that %sp must stay double-word aligned at
- all times, see ABI spec at http://www.sparc.org/standards/psABI3rd.pdf).
- To do this, we extend the *caller's* stack frame by 2 words and shift
- the output registers used for argument passing (%o0 - %o5, we are a *leaf*
- procedure because of the tail-jump) by 2 positions. This makes room in
- %o0 and %o1 for the additional arguments, namely hptr and a dummy (used
- for destination addr of jump on SPARC, return address on x86, ...). This
- shouldn't cause any problems for a C-like caller: alloca is implemented
- similarly, and local variables should be accessed via %fp, not %sp. In a
- nutshell: This should work! (Famous last words! :-)
- */
- adjustor = allocateExec(4*(11+1),&code);
- {
- unsigned long *const adj_code = (unsigned long *)adjustor;
-
- adj_code[ 0] = 0x9C23A008UL; /* sub %sp, 8, %sp */
- adj_code[ 1] = 0xDA23A060UL; /* st %o5, [%sp + 96] */
- adj_code[ 2] = 0xD823A05CUL; /* st %o4, [%sp + 92] */
- adj_code[ 3] = 0x9A10000BUL; /* mov %o3, %o5 */
- adj_code[ 4] = 0x9810000AUL; /* mov %o2, %o4 */
- adj_code[ 5] = 0x96100009UL; /* mov %o1, %o3 */
- adj_code[ 6] = 0x94100008UL; /* mov %o0, %o2 */
- adj_code[ 7] = 0x13000000UL; /* sethi %hi(wptr), %o1 */
- adj_code[ 7] |= ((unsigned long)wptr) >> 10;
- adj_code[ 8] = 0x11000000UL; /* sethi %hi(hptr), %o0 */
- adj_code[ 8] |= ((unsigned long)hptr) >> 10;
- adj_code[ 9] = 0x81C26000UL; /* jmp %o1 + %lo(wptr) */
- adj_code[ 9] |= ((unsigned long)wptr) & 0x000003FFUL;
- adj_code[10] = 0x90122000UL; /* or %o0, %lo(hptr), %o0 */
- adj_code[10] |= ((unsigned long)hptr) & 0x000003FFUL;
-
- adj_code[11] = (unsigned long)hptr;
-
- /* flush cache */
- asm("flush %0" : : "r" (adj_code ));
- asm("flush %0" : : "r" (adj_code + 2));
- asm("flush %0" : : "r" (adj_code + 4));
- asm("flush %0" : : "r" (adj_code + 6));
- asm("flush %0" : : "r" (adj_code + 10));
-
- /* max. 5 instructions latency, and we need at >= 1 for returning */
- asm("nop");
- asm("nop");
- asm("nop");
- asm("nop");
- }
-#elif defined(alpha_HOST_ARCH)
- /* Magic constant computed by inspecting the code length of
- the following assembly language snippet
- (offset and machine code prefixed; note that the machine code
- shown is longwords stored in little-endian order):
-
- <00>: 46520414 mov a2, a4
- <04>: 46100412 mov a0, a2
- <08>: a61b0020 ldq a0, 0x20(pv) # load up hptr
- <0c>: 46730415 mov a3, a5
- <10>: a77b0028 ldq pv, 0x28(pv) # load up wptr
- <14>: 46310413 mov a1, a3
- <18>: 6bfb---- jmp (pv), <hint> # jump to wptr (with hint)
- <1c>: 00000000 # padding for alignment
- <20>: [8 bytes for hptr quadword]
- <28>: [8 bytes for wptr quadword]
-
- The "computed" jump at <08> above is really a jump to a fixed
- location. Accordingly, we place an always-correct hint in the
- jump instruction, namely the address offset from <0c> to wptr,
- divided by 4, taking the lowest 14 bits.
-
- We only support passing 4 or fewer argument words, for the same
- reason described under sparc_HOST_ARCH above by JRS, 21 Aug 01.
- On the Alpha the first 6 integer arguments are in a0 through a5,
- and the rest on the stack. Hence we want to shuffle the original
- caller's arguments by two.
-
- On the Alpha the calling convention is so complex and dependent
- on the callee's signature -- for example, the stack pointer has
- to be a multiple of 16 -- that it seems impossible to me [ccshan]
- to handle the general case correctly without changing how the
- adjustor is called from C. For now, our solution of shuffling
- registers only and ignoring the stack only works if the original
- caller passed 4 or fewer argument words.
-
-TODO: Depending on how much allocation overhead stgMallocBytes uses for
- header information (more precisely, if the overhead is no more than
- 4 bytes), we should move the first three instructions above down by
- 4 bytes (getting rid of the nop), hence saving memory. [ccshan]
- */
- ASSERT(((StgWord64)wptr & 3) == 0);
- adjustor = allocateExec(48,&code);
- {
- StgWord64 *const code = (StgWord64 *)adjustor;
-
- code[0] = 0x4610041246520414L;
- code[1] = 0x46730415a61b0020L;
- code[2] = 0x46310413a77b0028L;
- code[3] = 0x000000006bfb0000L
- | (((StgWord32*)(wptr) - (StgWord32*)(code) - 3) & 0x3fff);
-
- code[4] = (StgWord64)hptr;
- code[5] = (StgWord64)wptr;
-
- /* Ensure that instruction cache is consistent with our new code */
- __asm__ volatile("call_pal %0" : : "i" (PAL_imb));
- }
-#elif defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS)
-
-#define OP_LO(op,lo) ((((unsigned)(op)) << 16) | (((unsigned)(lo)) & 0xFFFF))
-#define OP_HI(op,hi) ((((unsigned)(op)) << 16) | (((unsigned)(hi)) >> 16))
- {
- /* The PowerPC Linux (32-bit) calling convention is annoyingly complex.
- We need to calculate all the details of the stack frame layout,
- taking into account the types of all the arguments, and then
- generate code on the fly. */
-
- int src_gpr = 3, dst_gpr = 5;
- int fpr = 3;
- int src_offset = 0, dst_offset = 0;
- int n = strlen(typeString),i;
- int src_locs[n], dst_locs[n];
- int frameSize;
- unsigned *code;
-
- /* Step 1:
- Calculate where the arguments should go.
- src_locs[] will contain the locations of the arguments in the
- original stack frame passed to the adjustor.
- dst_locs[] will contain the locations of the arguments after the
- adjustor runs, on entry to the wrapper proc pointed to by wptr.
-
- This algorithm is based on the one described on page 3-19 of the
- System V ABI PowerPC Processor Supplement.
- */
- for(i=0;typeString[i];i++)
- {
- char t = typeString[i];
- if((t == 'f' || t == 'd') && fpr <= 8)
- src_locs[i] = dst_locs[i] = -32-(fpr++);
- else
+ // on 32-bit platforms, Double and Int64 occupy two words.
+ case 'd':
+ case 'l':
+ case 'L':
+ if(sizeof(void*) == 4)
{
- if((t == 'l' || t == 'L') && src_gpr <= 9)
- {
- if((src_gpr & 1) == 0)
- src_gpr++;
- src_locs[i] = -src_gpr;
- src_gpr += 2;
- }
- else if((t == 'w' || t == 'W') && src_gpr <= 10)
- {
- src_locs[i] = -(src_gpr++);
- }
- else
- {
- if(t == 'l' || t == 'L' || t == 'd')
- {
- if(src_offset % 8)
- src_offset += 4;
- }
- src_locs[i] = src_offset;
- src_offset += (t == 'l' || t == 'L' || t == 'd') ? 8 : 4;
- }
-
- if((t == 'l' || t == 'L') && dst_gpr <= 9)
- {
- if((dst_gpr & 1) == 0)
- dst_gpr++;
- dst_locs[i] = -dst_gpr;
- dst_gpr += 2;
- }
- else if((t == 'w' || t == 'W') && dst_gpr <= 10)
- {
- dst_locs[i] = -(dst_gpr++);
- }
- else
- {
- if(t == 'l' || t == 'L' || t == 'd')
- {
- if(dst_offset % 8)
- dst_offset += 4;
- }
- dst_locs[i] = dst_offset;
- dst_offset += (t == 'l' || t == 'L' || t == 'd') ? 8 : 4;
- }
+ sz += 2;
+ break;
}
- }
-
- frameSize = dst_offset + 8;
- frameSize = (frameSize+15) & ~0xF;
-
- /* Step 2:
- Build the adjustor.
- */
- // allocate space for at most 4 insns per parameter
- // plus 14 more instructions.
- adjustor = allocateExec(4 * (4*n + 14),&code);
- code = (unsigned*)adjustor;
-
- *code++ = 0x48000008; // b *+8
- // * Put the hptr in a place where freeHaskellFunctionPtr
- // can get at it.
- *code++ = (unsigned) hptr;
-
- // * save the link register
- *code++ = 0x7c0802a6; // mflr r0;
- *code++ = 0x90010004; // stw r0, 4(r1);
- // * and build a new stack frame
- *code++ = OP_LO(0x9421, -frameSize); // stwu r1, -frameSize(r1)
-
- // * now generate instructions to copy arguments
- // from the old stack frame into the new stack frame.
- for(i=n-1;i>=0;i--)
- {
- if(src_locs[i] < -32)
- ASSERT(dst_locs[i] == src_locs[i]);
- else if(src_locs[i] < 0)
- {
- // source in GPR.
- ASSERT(typeString[i] != 'f' && typeString[i] != 'd');
- if(dst_locs[i] < 0)
- {
- ASSERT(dst_locs[i] > -32);
- // dst is in GPR, too.
-
- if(typeString[i] == 'l' || typeString[i] == 'L')
- {
- // mr dst+1, src+1
- *code++ = 0x7c000378
- | ((-dst_locs[i]+1) << 16)
- | ((-src_locs[i]+1) << 11)
- | ((-src_locs[i]+1) << 21);
- }
- // mr dst, src
- *code++ = 0x7c000378
- | ((-dst_locs[i]) << 16)
- | ((-src_locs[i]) << 11)
- | ((-src_locs[i]) << 21);
- }
- else
- {
- if(typeString[i] == 'l' || typeString[i] == 'L')
- {
- // stw src+1, dst_offset+4(r1)
- *code++ = 0x90010000
- | ((-src_locs[i]+1) << 21)
- | (dst_locs[i] + 4);
- }
-
- // stw src, dst_offset(r1)
- *code++ = 0x90010000
- | ((-src_locs[i]) << 21)
- | (dst_locs[i] + 8);
- }
- }
- else
- {
- ASSERT(dst_locs[i] >= 0);
- ASSERT(typeString[i] != 'f' && typeString[i] != 'd');
-
- if(typeString[i] == 'l' || typeString[i] == 'L')
- {
- // lwz r0, src_offset(r1)
- *code++ = 0x80010000
- | (src_locs[i] + frameSize + 8 + 4);
- // stw r0, dst_offset(r1)
- *code++ = 0x90010000
- | (dst_locs[i] + 8 + 4);
- }
- // lwz r0, src_offset(r1)
- *code++ = 0x80010000
- | (src_locs[i] + frameSize + 8);
- // stw r0, dst_offset(r1)
- *code++ = 0x90010000
- | (dst_locs[i] + 8);
- }
- }
-
- // * hptr will be the new first argument.
- // lis r3, hi(hptr)
- *code++ = OP_HI(0x3c60, hptr);
- // ori r3,r3,lo(hptr)
- *code++ = OP_LO(0x6063, hptr);
-
- // * we need to return to a piece of code
- // which will tear down the stack frame.
- // lis r11,hi(obscure_ccall_ret_code)
- *code++ = OP_HI(0x3d60, obscure_ccall_ret_code);
- // ori r11,r11,lo(obscure_ccall_ret_code)
- *code++ = OP_LO(0x616b, obscure_ccall_ret_code);
- // mtlr r11
- *code++ = 0x7d6803a6;
-
- // * jump to wptr
- // lis r11,hi(wptr)
- *code++ = OP_HI(0x3d60, wptr);
- // ori r11,r11,lo(wptr)
- *code++ = OP_LO(0x616b, wptr);
- // mtctr r11
- *code++ = 0x7d6903a6;
- // bctr
- *code++ = 0x4e800420;
-
- // Flush the Instruction cache:
- {
- unsigned *p = adjustor;
- while(p < code)
- {
- __asm__ volatile ("dcbf 0,%0\n\tsync\n\ticbi 0,%0"
- : : "r" (p));
- p++;
- }
- __asm__ volatile ("sync\n\tisync");
+ /* fall through */
+ // everything else is one word.
+ default:
+ sz += 1;
}
}
-
-#elif defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH)
-
-#define OP_LO(op,lo) ((((unsigned)(op)) << 16) | (((unsigned)(lo)) & 0xFFFF))
-#define OP_HI(op,hi) ((((unsigned)(op)) << 16) | (((unsigned)(hi)) >> 16))
- {
- /* The following code applies to all PowerPC and PowerPC64 platforms
- whose stack layout is based on the AIX ABI.
-
- Besides (obviously) AIX, this includes
- Mac OS 9 and BeOS/PPC and Mac OS X PPC (may they rest in peace),
- which use the 32-bit AIX ABI
- powerpc64-linux,
- which uses the 64-bit AIX ABI.
-
- The actual stack-frame shuffling is implemented out-of-line
- in the function adjustorCode, in AdjustorAsm.S.
- Here, we set up an AdjustorStub structure, which
- is a function descriptor with a pointer to the AdjustorStub
- struct in the position of the TOC that is loaded
- into register r2.
-
- One nice thing about this is that there is _no_ code generated at
- runtime on the platforms that have function descriptors.
- */
- AdjustorStub *adjustorStub;
- int sz = 0, extra_sz, total_sz;
-
-#if defined(FUNDESCS)
- adjustorStub = stgMallocBytes(sizeof(AdjustorStub), "createAdjustor");
-#else
- adjustorStub = allocateExec(sizeof(AdjustorStub),&code);
-#endif
- adjustor = adjustorStub;
-
- adjustorStub->code = (void*) &adjustorCode;
-
-#if defined(FUNDESCS)
- // function descriptors are a cool idea.
- // We don't need to generate any code at runtime.
- adjustorStub->toc = adjustorStub;
-#else
-
- // no function descriptors :-(
- // We need to do things "by hand".
-#if defined(powerpc_HOST_ARCH)
- // lis r2, hi(adjustorStub)
- adjustorStub->lis = OP_HI(0x3c40, adjustorStub);
- // ori r2, r2, lo(adjustorStub)
- adjustorStub->ori = OP_LO(0x6042, adjustorStub);
- // lwz r0, code(r2)
- adjustorStub->lwz = OP_LO(0x8002, (char*)(&adjustorStub->code)
- - (char*)adjustorStub);
- // mtctr r0
- adjustorStub->mtctr = 0x7c0903a6;
- // bctr
- adjustorStub->bctr = 0x4e800420;
-#else
- barf("adjustor creation not supported on this platform");
-#endif
-
- // Flush the Instruction cache:
- {
- int n = sizeof(AdjustorStub)/sizeof(unsigned);
- unsigned *p = (unsigned*)adjustor;
- while(n--)
- {
- __asm__ volatile ("dcbf 0,%0\n\tsync\n\ticbi 0,%0"
- : : "r" (p));
- p++;
- }
- __asm__ volatile ("sync\n\tisync");
- }
-#endif
-
- // Calculate the size of the stack frame, in words.
- sz = totalArgumentSize(typeString);
-
- // The first eight words of the parameter area
- // are just "backing store" for the parameters passed in
- // the GPRs. extra_sz is the number of words beyond those first
- // 8 words.
- extra_sz = sz - 8;
- if(extra_sz < 0)
- extra_sz = 0;
-
- // Calculate the total size of the stack frame.
- total_sz = (6 /* linkage area */
- + 8 /* minimum parameter area */
- + 2 /* two extra arguments */
- + extra_sz)*sizeof(StgWord);
-
- // align to 16 bytes.
- // AIX only requires 8 bytes, but who cares?
- total_sz = (total_sz+15) & ~0xF;
-
- // Fill in the information that adjustorCode in AdjustorAsm.S
- // will use to create a new stack frame with the additional args.
- adjustorStub->hptr = hptr;
- adjustorStub->wptr = wptr;
- adjustorStub->negative_framesize = -total_sz;
- adjustorStub->extrawords_plus_one = extra_sz + 1;
- }
-
-#elif defined(ia64_HOST_ARCH)
-/*
- Up to 8 inputs are passed in registers. We flush the last two inputs to
- the stack, initially into the 16-byte scratch region left by the caller.
- We then shuffle the others along by 4 (taking 2 registers for ourselves
- to save return address and previous function state - we need to come back
- here on the way out to restore the stack, so this is a real function
- rather than just a trampoline).
-
- The function descriptor we create contains the gp of the target function
- so gp is already loaded correctly.
-
- [MLX] alloc r16=ar.pfs,10,2,0
- movl r17=wptr
- [MII] st8.spill [r12]=r38,8 // spill in6 (out4)
- mov r41=r37 // out7 = in5 (out3)
- mov r40=r36;; // out6 = in4 (out2)
- [MII] st8.spill [r12]=r39 // spill in7 (out5)
- mov.sptk b6=r17,50
- mov r38=r34;; // out4 = in2 (out0)
- [MII] mov r39=r35 // out5 = in3 (out1)
- mov r37=r33 // out3 = in1 (loc1)
- mov r36=r32 // out2 = in0 (loc0)
- [MLX] adds r12=-24,r12 // update sp
- movl r34=hptr;; // out0 = hptr
- [MIB] mov r33=r16 // loc1 = ar.pfs
- mov r32=b0 // loc0 = retaddr
- br.call.sptk.many b0=b6;;
-
- [MII] adds r12=-16,r12
- mov b0=r32
- mov.i ar.pfs=r33
- [MFB] nop.m 0x0
- nop.f 0x0
- br.ret.sptk.many b0;;
-*/
-
-/* These macros distribute a long constant into the two words of an MLX bundle */
-#define BITS(val,start,count) (((val) >> (start)) & ((1 << (count))-1))
-#define MOVL_LOWORD(val) (BITS(val,22,18) << 46)
-#define MOVL_HIWORD(val) ( (BITS(val,0,7) << 36) \
- | (BITS(val,7,9) << 50) \
- | (BITS(val,16,5) << 45) \
- | (BITS(val,21,1) << 44) \
- | (BITS(val,40,23)) \
- | (BITS(val,63,1) << 59))
-
- {
- StgStablePtr stable;
- IA64FunDesc *wdesc = (IA64FunDesc *)wptr;
- StgWord64 wcode = wdesc->ip;
- IA64FunDesc *fdesc;
- StgWord64 *code;
-
- /* we allocate on the Haskell heap since malloc'd memory isn't
- * executable - argh */
- /* Allocated memory is word-aligned (8 bytes) but functions on ia64
- * must be aligned to 16 bytes. We allocate an extra 8 bytes of
- * wiggle room so that we can put the code on a 16 byte boundary. */
- adjustor = stgAllocStable(sizeof(IA64FunDesc)+18*8+8, &stable);
-
- fdesc = (IA64FunDesc *)adjustor;
- code = (StgWord64 *)(fdesc + 1);
- /* add 8 bytes to code if needed to align to a 16-byte boundary */
- if ((StgWord64)code & 15) code++;
- fdesc->ip = (StgWord64)code;
- fdesc->gp = wdesc->gp;
-
- code[0] = 0x0000058004288004 | MOVL_LOWORD(wcode);
- code[1] = 0x6000000220000000 | MOVL_HIWORD(wcode);
- code[2] = 0x029015d818984001;
- code[3] = 0x8401200500420094;
- code[4] = 0x886011d8189c0001;
- code[5] = 0x84011004c00380c0;
- code[6] = 0x0250210046013800;
- code[7] = 0x8401000480420084;
- code[8] = 0x0000233f19a06005 | MOVL_LOWORD((StgWord64)hptr);
- code[9] = 0x6000000440000000 | MOVL_HIWORD((StgWord64)hptr);
- code[10] = 0x0200210020010811;
- code[11] = 0x1080006800006200;
- code[12] = 0x0000210018406000;
- code[13] = 0x00aa021000038005;
- code[14] = 0x000000010000001d;
- code[15] = 0x0084000880000200;
-
- /* save stable pointers in convenient form */
- code[16] = (StgWord64)hptr;
- code[17] = (StgWord64)stable;
- }
-#else
- barf("adjustor creation not supported on this platform");
-#endif
- break;
-
- default:
- ASSERT(0);
- break;
- }
-
- /* Have fun! */
- return code;
-}
-
-
-void
-freeHaskellFunctionPtr(void* ptr)
-{
-#if defined(i386_HOST_ARCH)
- if ( *(unsigned char*)ptr != 0xe8 &&
- *(unsigned char*)ptr != 0x58 ) {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
- if (*(unsigned char*)ptr == 0xe8) { /* Aha, a ccall adjustor! */
- freeStablePtr(((AdjustorStub*)ptr)->hptr);
- } else {
- freeStablePtr(*((StgStablePtr*)((unsigned char*)ptr + 0x02)));
- }
-#elif defined(x86_64_HOST_ARCH)
- if ( *(StgWord16 *)ptr == 0x894d ) {
- freeStablePtr(*(StgStablePtr*)((StgWord8*)ptr+
-#if defined(mingw32_HOST_OS)
- 0x28
-#else
- 0x20
-#endif
- ));
-#if !defined(mingw32_HOST_OS)
- } else if ( *(StgWord16 *)ptr == 0x5141 ) {
- freeStablePtr(*(StgStablePtr*)((StgWord8*)ptr+0x30));
-#endif
-#if defined(mingw32_HOST_OS)
- } else if ( *(StgWord16 *)ptr == 0x8348 ) {
- freeStablePtr(*(StgStablePtr*)((StgWord8*)ptr+0x48));
-#endif
- } else {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
-#elif defined(sparc_HOST_ARCH)
- if ( *(unsigned long*)ptr != 0x9C23A008UL ) {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
-
- /* Free the stable pointer first..*/
- freeStablePtr(*((StgStablePtr*)((unsigned long*)ptr + 11)));
-#elif defined(alpha_HOST_ARCH)
- if ( *(StgWord64*)ptr != 0xa77b0018a61b0010L ) {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
-
- /* Free the stable pointer first..*/
- freeStablePtr(*((StgStablePtr*)((unsigned char*)ptr + 0x10)));
-#elif defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS)
- if ( *(StgWord*)ptr != 0x48000008 ) {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
- freeStablePtr(((StgStablePtr*)ptr)[1]);
-#elif defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH)
- if ( ((AdjustorStub*)ptr)->code != (StgFunPtr) &adjustorCode ) {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
- freeStablePtr(((AdjustorStub*)ptr)->hptr);
-#elif defined(ia64_HOST_ARCH)
- IA64FunDesc *fdesc = (IA64FunDesc *)ptr;
- StgWord64 *code = (StgWord64 *)(fdesc+1);
-
- if (fdesc->ip != (StgWord64)code) {
- errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
- return;
- }
- freeStablePtr((StgStablePtr)code[16]);
- freeStablePtr((StgStablePtr)code[17]);
- return;
-#else
- ASSERT(0);
-#endif
- // Can't write to this memory, it is only executable:
- // *((unsigned char*)ptr) = '\0';
-
- freeExec(ptr);
+ return sz;
}
-
-#endif // !USE_LIBFFI_FOR_ADJUSTORS
diff --git a/rts/Adjustor.h b/rts/Adjustor.h
new file mode 100644
index 0000000000..a6dcac6d56
--- /dev/null
+++ b/rts/Adjustor.h
@@ -0,0 +1,7 @@
+/* -----------------------------------------------------------------------------
+ * Common utilities used in adjustor implementations
+ * ---------------------------------------------------------------------------*/
+
+#pragma once
+
+int totalArgumentSize(const char *typeString);
diff --git a/rts/adjustor/LibffiAdjustor.c b/rts/adjustor/LibffiAdjustor.c
new file mode 100644
index 0000000000..4b843c1b82
--- /dev/null
+++ b/rts/adjustor/LibffiAdjustor.c
@@ -0,0 +1,176 @@
+/* -----------------------------------------------------------------------------
+ * libffi-based adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+#include "sm/Storage.h"
+#include "Hash.h"
+
+#include "ffi.h"
+#include <string.h>
+
+/* Maps AdjustorExecutable* to AdjustorWritable*. */
+static HashTable* allocatedExecs;
+
+static AdjustorWritable allocate_adjustor(AdjustorExecutable *exec_ret)
+{
+ AdjustorWritable writ;
+ ffi_closure* cl;
+
+ ACQUIRE_SM_LOCK;
+ cl = writ = ffi_closure_alloc(sizeof(ffi_closure), exec_ret);
+ if (cl != NULL) {
+ if (allocatedExecs == NULL) {
+ allocatedExecs = allocHashTable();
+ }
+ insertHashTable(allocatedExecs, (StgWord)*exec_ret, writ);
+ }
+ RELEASE_SM_LOCK;
+ return writ;
+}
+
+static AdjustorWritable exec_to_writable(AdjustorExecutable exec)
+{
+ AdjustorWritable writ;
+ ACQUIRE_SM_LOCK;
+ if (allocatedExecs == NULL ||
+ (writ = lookupHashTable(allocatedExecs, (StgWord)exec)) == NULL) {
+ RELEASE_SM_LOCK;
+ barf("exec_to_writable: not found");
+ }
+ RELEASE_SM_LOCK;
+ return writ;
+}
+
+static void free_adjustor(AdjustorExecutable exec)
+{
+ AdjustorWritable writ;
+ ffi_closure* cl;
+ cl = writ = exec_to_writable(exec);
+ ACQUIRE_SM_LOCK;
+ removeHashTable(allocatedExecs, (StgWord)exec, writ);
+ ffi_closure_free(cl);
+ RELEASE_SM_LOCK;
+}
+
+
+/* Note [Freeing libffi adjustors]
+ * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ *
+ * HOW ADJUSTORS/CLOSURES WORK ON LIBFFI:
+ * libffi's ffi_closure_alloc() function gives you two pointers to a closure,
+ * 1. the writable pointer, and 2. the executable pointer. You write the
+ * closure into the writable pointer (and ffi_prep_closure_loc() will do this
+ * for you) and you execute it at the executable pointer.
+ *
+ * THE PROBLEM:
+ * The RTS deals only with the executable pointer, but when it comes time to
+ * free the closure, libffi wants the writable pointer back that it gave you
+ * when you allocated it.
+ *
+ * On Linux we used to solve this problem by storing the address of the writable
+ * mapping into itself, then returning both writable and executable pointers
+ * plus 1 machine word for preparing the closure for use by the RTS (see the
+ * Linux version of allocateExec() in rts/sm/Storage.c). When we want to recover
+ * the writable address, we subtract 1 word from the executable address and
+ * fetch. This works because Linux kernel magic gives us two pointers with
+ * different addresses that refer to the same memory. Whatever you write into
+ * the writable address can be read back at the executable address. This method
+ * is very efficient.
+ *
+ * On iOS this breaks for two reasons: 1. the two pointers do not refer to
+ * the same memory (so we can't retrieve anything stored into the writable
+ * pointer if we only have the exec pointer), and 2. libffi's
+ * ffi_closure_alloc() assumes the pointer it has returned you is a
+ * ffi_closure structure and treats it as such: It uses that memory to
+ * communicate with ffi_prep_closure_loc(). On Linux by contrast
+ * ffi_closure_alloc() is viewed simply as a memory allocation, and only
+ * ffi_prep_closure_loc() deals in ffi_closure structures. Each of these
+ * differences is enough make the efficient way used on Linux not work on iOS.
+ * Instead on iOS we use hash tables to recover the writable address from the
+ * executable one. This method is conservative and would almost certainly work
+ * on any platform. This is what we now do everywhere.
+ */
+void
+freeHaskellFunctionPtr(void* ptr)
+{
+ ffi_closure *cl;
+
+ cl = exec_to_writable(ptr);
+ freeStablePtr(cl->user_data);
+ stgFree(cl->cif->arg_types);
+ stgFree(cl->cif);
+ free_adjustor(ptr);
+}
+
+static ffi_type * char_to_ffi_type(char c)
+{
+ switch (c) {
+ case 'v': return &ffi_type_void;
+ case 'f': return &ffi_type_float;
+ case 'd': return &ffi_type_double;
+ case 'L': return &ffi_type_sint64;
+ case 'l': return &ffi_type_uint64;
+ case 'W': return &ffi_type_sint32;
+ case 'w': return &ffi_type_uint32;
+ case 'S': return &ffi_type_sint16;
+ case 's': return &ffi_type_uint16;
+ case 'B': return &ffi_type_sint8;
+ case 'b': return &ffi_type_uint8;
+ case 'p': return &ffi_type_pointer;
+ default: barf("char_to_ffi_type: unknown type '%c'", c);
+ }
+}
+
+void*
+createAdjustor (int cconv,
+ StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString)
+{
+ ffi_cif *cif;
+ ffi_type **arg_types;
+ uint32_t n_args, i;
+ ffi_type *result_type;
+ ffi_closure *cl;
+ int r, abi;
+ void *code;
+
+ n_args = strlen(typeString) - 1;
+ cif = stgMallocBytes(sizeof(ffi_cif), "createAdjustor");
+ arg_types = stgMallocBytes(n_args * sizeof(ffi_type*), "createAdjustor");
+
+ result_type = char_to_ffi_type(typeString[0]);
+ for (i=0; i < n_args; i++) {
+ arg_types[i] = char_to_ffi_type(typeString[i+1]);
+ }
+ switch (cconv) {
+#if defined(mingw32_HOST_OS) && defined(i386_HOST_ARCH)
+ case 0: /* stdcall */
+ abi = FFI_STDCALL;
+ break;
+#endif /* defined(mingw32_HOST_OS) && defined(i386_HOST_ARCH) */
+ case 1: /* ccall */
+ abi = FFI_DEFAULT_ABI;
+ break;
+ default:
+ barf("createAdjustor: convention %d not supported on this platform", cconv);
+ }
+
+ r = ffi_prep_cif(cif, abi, n_args, result_type, arg_types);
+ if (r != FFI_OK) barf("ffi_prep_cif failed: %d", r);
+
+ cl = allocate_adjustor(&code);
+ if (cl == NULL) {
+ barf("createAdjustor: failed to allocate memory");
+ }
+
+ r = ffi_prep_closure_loc(cl, cif, (void*)wptr, hptr/*userdata*/, code);
+ if (r != FFI_OK) barf("ffi_prep_closure_loc failed: %d", r);
+
+ return (void*)code;
+}
diff --git a/rts/adjustor/NativeAlpha.c b/rts/adjustor/NativeAlpha.c
new file mode 100644
index 0000000000..7c69b5d3b7
--- /dev/null
+++ b/rts/adjustor/NativeAlpha.c
@@ -0,0 +1,110 @@
+/* -----------------------------------------------------------------------------
+ * Alpha architecture adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+
+/* To get the definition of PAL_imb: */
+#if defined(linux_HOST_OS)
+# include <asm/pal.h>
+#else
+# include <machine/pal.h>
+#endif
+
+void*
+createAdjustor(int cconv, StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString STG_UNUSED
+ )
+{
+ void *adjustor = NULL;
+ void *code = NULL;
+
+ switch (cconv)
+ {
+ case 1: /* _ccall */
+ /* Magic constant computed by inspecting the code length of
+ the following assembly language snippet
+ (offset and machine code prefixed; note that the machine code
+ shown is longwords stored in little-endian order):
+
+ <00>: 46520414 mov a2, a4
+ <04>: 46100412 mov a0, a2
+ <08>: a61b0020 ldq a0, 0x20(pv) # load up hptr
+ <0c>: 46730415 mov a3, a5
+ <10>: a77b0028 ldq pv, 0x28(pv) # load up wptr
+ <14>: 46310413 mov a1, a3
+ <18>: 6bfb---- jmp (pv), <hint> # jump to wptr (with hint)
+ <1c>: 00000000 # padding for alignment
+ <20>: [8 bytes for hptr quadword]
+ <28>: [8 bytes for wptr quadword]
+
+ The "computed" jump at <08> above is really a jump to a fixed
+ location. Accordingly, we place an always-correct hint in the
+ jump instruction, namely the address offset from <0c> to wptr,
+ divided by 4, taking the lowest 14 bits.
+
+ We only support passing 4 or fewer argument words, for the same
+ reason described under sparc_HOST_ARCH above by JRS, 21 Aug 01.
+ On the Alpha the first 6 integer arguments are in a0 through a5,
+ and the rest on the stack. Hence we want to shuffle the original
+ caller's arguments by two.
+
+ On the Alpha the calling convention is so complex and dependent
+ on the callee's signature -- for example, the stack pointer has
+ to be a multiple of 16 -- that it seems impossible to me [ccshan]
+ to handle the general case correctly without changing how the
+ adjustor is called from C. For now, our solution of shuffling
+ registers only and ignoring the stack only works if the original
+ caller passed 4 or fewer argument words.
+
+TODO: Depending on how much allocation overhead stgMallocBytes uses for
+ header information (more precisely, if the overhead is no more than
+ 4 bytes), we should move the first three instructions above down by
+ 4 bytes (getting rid of the nop), hence saving memory. [ccshan]
+ */
+ ASSERT(((StgWord64)wptr & 3) == 0);
+ adjustor = allocateExec(48,&code);
+ {
+ StgWord64 *const code = (StgWord64 *)adjustor;
+
+ code[0] = 0x4610041246520414L;
+ code[1] = 0x46730415a61b0020L;
+ code[2] = 0x46310413a77b0028L;
+ code[3] = 0x000000006bfb0000L
+ | (((StgWord32*)(wptr) - (StgWord32*)(code) - 3) & 0x3fff);
+
+ code[4] = (StgWord64)hptr;
+ code[5] = (StgWord64)wptr;
+
+ /* Ensure that instruction cache is consistent with our new code */
+ __asm__ volatile("call_pal %0" : : "i" (PAL_imb));
+ }
+
+ default:
+ barf("createAdjustor: Unsupported calling convention");
+ }
+
+ return code;
+}
+
+void
+freeHaskellFunctionPtr(void* ptr)
+{
+ if ( *(StgWord64*)ptr != 0xa77b0018a61b0010L ) {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+
+ /* Free the stable pointer first..*/
+ freeStablePtr(*((StgStablePtr*)((unsigned char*)ptr + 0x10)));
+
+ // Can't write to this memory, it is only executable:
+ // *((unsigned char*)ptr) = '\0';
+
+ freeExec(ptr);
+}
diff --git a/rts/adjustor/NativeAmd64.c b/rts/adjustor/NativeAmd64.c
new file mode 100644
index 0000000000..fae215abe8
--- /dev/null
+++ b/rts/adjustor/NativeAmd64.c
@@ -0,0 +1,339 @@
+/* -----------------------------------------------------------------------------
+ * AMD64 architecture adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+
+#if defined(LEADING_UNDERSCORE)
+#define UNDERSCORE "_"
+#else
+#define UNDERSCORE ""
+#endif
+
+/*
+ Now here's something obscure for you:
+
+ When generating an adjustor thunk that uses the C calling
+ convention, we have to make sure that the thunk kicks off
+ the process of jumping into Haskell with a tail jump. Why?
+ Because as a result of jumping in into Haskell we may end
+ up freeing the very adjustor thunk we came from using
+ freeHaskellFunctionPtr(). Hence, we better not return to
+ the adjustor code on our way out, since it could by then
+ point to junk.
+
+ The fix is readily at hand, just include the opcodes
+ for the C stack fixup code that we need to perform when
+ returning in some static piece of memory and arrange
+ to return to it before tail jumping from the adjustor thunk.
+*/
+static void GNUC3_ATTRIBUTE(used) obscure_ccall_wrapper(void)
+{
+ __asm__ (
+ ".globl " UNDERSCORE "obscure_ccall_ret_code\n"
+ UNDERSCORE "obscure_ccall_ret_code:\n\t"
+ "addq $0x8, %rsp\n\t"
+#if defined(mingw32_HOST_OS)
+ /* On Win64, we had to put the original return address after the
+ arg 1-4 spill slots, ro now we have to move it back */
+ "movq 0x20(%rsp), %rcx\n"
+ "movq %rcx, (%rsp)\n"
+#endif /* defined(mingw32_HOST_OS) */
+ "ret"
+ );
+}
+
+extern void obscure_ccall_ret_code(void);
+
+void*
+createAdjustor(int cconv, StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString
+ )
+{
+ void *adjustor = NULL;
+ void *code = NULL;
+
+ switch (cconv)
+ {
+ case 1: /* _ccall */
+#if defined(mingw32_HOST_OS)
+ /*
+ stack at call:
+ argn
+ ...
+ arg5
+ return address
+ %rcx,%rdx,%r8,%r9 = arg1..arg4
+
+ if there are <4 integer args, then we can just push the
+ StablePtr into %rcx and shuffle the other args up.
+
+ If there are >=4 integer args, then we have to flush one arg
+ to the stack, and arrange to adjust the stack ptr on return.
+ The stack will be rearranged to this:
+
+ argn
+ ...
+ arg5
+ return address *** <-- dummy arg in stub fn.
+ arg4
+ obscure_ccall_ret_code
+
+ This unfortunately means that the type of the stub function
+ must have a dummy argument for the original return address
+ pointer inserted just after the 4th integer argument.
+
+ Code for the simple case:
+
+ 0: 4d 89 c1 mov %r8,%r9
+ 3: 49 89 d0 mov %rdx,%r8
+ 6: 48 89 ca mov %rcx,%rdx
+ 9: f2 0f 10 da movsd %xmm2,%xmm3
+ d: f2 0f 10 d1 movsd %xmm1,%xmm2
+ 11: f2 0f 10 c8 movsd %xmm0,%xmm1
+ 15: 48 8b 0d 0c 00 00 00 mov 0xc(%rip),%rcx # 28 <.text+0x28>
+ 1c: ff 25 0e 00 00 00 jmpq *0xe(%rip) # 30 <.text+0x30>
+ 22: 90 nop
+ [...]
+
+
+ And the version for >=4 integer arguments:
+
+[we want to push the 4th argument (either %r9 or %xmm3, depending on
+ whether it is a floating arg or not) and the return address onto the
+ stack. However, slots 1-4 are reserved for code we call to spill its
+ args 1-4 into, so we can't just push them onto the bottom of the stack.
+ So first put the 4th argument onto the stack, above what will be the
+ spill slots.]
+ 0: 48 83 ec 08 sub $0x8,%rsp
+[if non-floating arg, then do this:]
+ 4: 90 nop
+ 5: 4c 89 4c 24 20 mov %r9,0x20(%rsp)
+[else if floating arg then do this:]
+ 4: f2 0f 11 5c 24 20 movsd %xmm3,0x20(%rsp)
+[end if]
+[Now push the new return address onto the stack]
+ a: ff 35 30 00 00 00 pushq 0x30(%rip) # 40 <.text+0x40>
+[But the old return address has been moved up into a spill slot, so
+ we need to move it above them]
+ 10: 4c 8b 4c 24 10 mov 0x10(%rsp),%r9
+ 15: 4c 89 4c 24 30 mov %r9,0x30(%rsp)
+[Now we do the normal register shuffle-up etc]
+ 1a: 4d 89 c1 mov %r8,%r9
+ 1d: 49 89 d0 mov %rdx,%r8
+ 20: 48 89 ca mov %rcx,%rdx
+ 23: f2 0f 10 da movsd %xmm2,%xmm3
+ 27: f2 0f 10 d1 movsd %xmm1,%xmm2
+ 2b: f2 0f 10 c8 movsd %xmm0,%xmm1
+ 2f: 48 8b 0d 12 00 00 00 mov 0x12(%rip),%rcx # 48 <.text+0x48>
+ 36: ff 25 14 00 00 00 jmpq *0x14(%rip) # 50 <.text+0x50>
+ 3c: 90 nop
+ 3d: 90 nop
+ 3e: 90 nop
+ 3f: 90 nop
+ [...]
+
+ */
+ {
+ StgWord8 *adj_code;
+
+ // determine whether we have 4 or more integer arguments,
+ // and therefore need to flush one to the stack.
+ if ((typeString[0] == '\0') ||
+ (typeString[1] == '\0') ||
+ (typeString[2] == '\0') ||
+ (typeString[3] == '\0')) {
+
+ adjustor = allocateExec(0x38,&code);
+ adj_code = (StgWord8*)adjustor;
+
+ *(StgInt32 *)adj_code = 0x49c1894d;
+ *(StgInt32 *)(adj_code+0x4) = 0x8948d089;
+ *(StgInt32 *)(adj_code+0x8) = 0x100ff2ca;
+ *(StgInt32 *)(adj_code+0xc) = 0x100ff2da;
+ *(StgInt32 *)(adj_code+0x10) = 0x100ff2d1;
+ *(StgInt32 *)(adj_code+0x14) = 0x0d8b48c8;
+ *(StgInt32 *)(adj_code+0x18) = 0x0000000c;
+
+ *(StgInt32 *)(adj_code+0x1c) = 0x000e25ff;
+ *(StgInt32 *)(adj_code+0x20) = 0x00000000;
+ *(StgInt64 *)(adj_code+0x28) = (StgInt64)hptr;
+ *(StgInt64 *)(adj_code+0x30) = (StgInt64)wptr;
+ }
+ else
+ {
+ int fourthFloating;
+
+ fourthFloating = (typeString[3] == 'f' || typeString[3] == 'd');
+ adjustor = allocateExec(0x58,&code);
+ adj_code = (StgWord8*)adjustor;
+ *(StgInt32 *)adj_code = 0x08ec8348;
+ *(StgInt32 *)(adj_code+0x4) = fourthFloating ? 0x5c110ff2
+ : 0x4c894c90;
+ *(StgInt32 *)(adj_code+0x8) = 0x35ff2024;
+ *(StgInt32 *)(adj_code+0xc) = 0x00000030;
+ *(StgInt32 *)(adj_code+0x10) = 0x244c8b4c;
+ *(StgInt32 *)(adj_code+0x14) = 0x4c894c10;
+ *(StgInt32 *)(adj_code+0x18) = 0x894d3024;
+ *(StgInt32 *)(adj_code+0x1c) = 0xd08949c1;
+ *(StgInt32 *)(adj_code+0x20) = 0xf2ca8948;
+ *(StgInt32 *)(adj_code+0x24) = 0xf2da100f;
+ *(StgInt32 *)(adj_code+0x28) = 0xf2d1100f;
+ *(StgInt32 *)(adj_code+0x2c) = 0x48c8100f;
+ *(StgInt32 *)(adj_code+0x30) = 0x00120d8b;
+ *(StgInt32 *)(adj_code+0x34) = 0x25ff0000;
+ *(StgInt32 *)(adj_code+0x38) = 0x00000014;
+ *(StgInt32 *)(adj_code+0x3c) = 0x90909090;
+ *(StgInt64 *)(adj_code+0x40) = (StgInt64)obscure_ccall_ret_code;
+ *(StgInt64 *)(adj_code+0x48) = (StgInt64)hptr;
+ *(StgInt64 *)(adj_code+0x50) = (StgInt64)wptr;
+ }
+ }
+
+# else
+ /*
+ stack at call:
+ argn
+ ...
+ arg7
+ return address
+ %rdi,%rsi,%rdx,%rcx,%r8,%r9 = arg1..arg6
+
+ if there are <6 integer args, then we can just push the
+ StablePtr into %edi and shuffle the other args up.
+
+ If there are >=6 integer args, then we have to flush one arg
+ to the stack, and arrange to adjust the stack ptr on return.
+ The stack will be rearranged to this:
+
+ argn
+ ...
+ arg7
+ return address *** <-- dummy arg in stub fn.
+ arg6
+ obscure_ccall_ret_code
+
+ This unfortunately means that the type of the stub function
+ must have a dummy argument for the original return address
+ pointer inserted just after the 6th integer argument.
+
+ Code for the simple case:
+
+ 0: 4d 89 c1 mov %r8,%r9
+ 3: 49 89 c8 mov %rcx,%r8
+ 6: 48 89 d1 mov %rdx,%rcx
+ 9: 48 89 f2 mov %rsi,%rdx
+ c: 48 89 fe mov %rdi,%rsi
+ f: 48 8b 3d 0a 00 00 00 mov 10(%rip),%rdi
+ 16: ff 25 0c 00 00 00 jmpq *12(%rip)
+ ...
+ 20: .quad 0 # aligned on 8-byte boundary
+ 28: .quad 0 # aligned on 8-byte boundary
+
+
+ And the version for >=6 integer arguments:
+
+ 0: 41 51 push %r9
+ 2: ff 35 20 00 00 00 pushq 32(%rip) # 28 <ccall_adjustor+0x28>
+ 8: 4d 89 c1 mov %r8,%r9
+ b: 49 89 c8 mov %rcx,%r8
+ e: 48 89 d1 mov %rdx,%rcx
+ 11: 48 89 f2 mov %rsi,%rdx
+ 14: 48 89 fe mov %rdi,%rsi
+ 17: 48 8b 3d 12 00 00 00 mov 18(%rip),%rdi # 30 <ccall_adjustor+0x30>
+ 1e: ff 25 14 00 00 00 jmpq *20(%rip) # 38 <ccall_adjustor+0x38>
+ ...
+ 28: .quad 0 # aligned on 8-byte boundary
+ 30: .quad 0 # aligned on 8-byte boundary
+ 38: .quad 0 # aligned on 8-byte boundary
+ */
+
+ {
+ int i = 0;
+ char *c;
+ StgWord8 *adj_code;
+
+ // determine whether we have 6 or more integer arguments,
+ // and therefore need to flush one to the stack.
+ for (c = typeString; *c != '\0'; c++) {
+ if (*c != 'f' && *c != 'd') i++;
+ if (i == 6) break;
+ }
+
+ if (i < 6) {
+ adjustor = allocateExec(0x30,&code);
+ adj_code = (StgWord8*)adjustor;
+
+ *(StgInt32 *)adj_code = 0x49c1894d;
+ *(StgInt32 *)(adj_code+0x4) = 0x8948c889;
+ *(StgInt32 *)(adj_code+0x8) = 0xf28948d1;
+ *(StgInt32 *)(adj_code+0xc) = 0x48fe8948;
+ *(StgInt32 *)(adj_code+0x10) = 0x000a3d8b;
+ *(StgInt32 *)(adj_code+0x14) = 0x25ff0000;
+ *(StgInt32 *)(adj_code+0x18) = 0x0000000c;
+ *(StgInt64 *)(adj_code+0x20) = (StgInt64)hptr;
+ *(StgInt64 *)(adj_code+0x28) = (StgInt64)wptr;
+ }
+ else
+ {
+ adjustor = allocateExec(0x40,&code);
+ adj_code = (StgWord8*)adjustor;
+
+ *(StgInt32 *)adj_code = 0x35ff5141;
+ *(StgInt32 *)(adj_code+0x4) = 0x00000020;
+ *(StgInt32 *)(adj_code+0x8) = 0x49c1894d;
+ *(StgInt32 *)(adj_code+0xc) = 0x8948c889;
+ *(StgInt32 *)(adj_code+0x10) = 0xf28948d1;
+ *(StgInt32 *)(adj_code+0x14) = 0x48fe8948;
+ *(StgInt32 *)(adj_code+0x18) = 0x00123d8b;
+ *(StgInt32 *)(adj_code+0x1c) = 0x25ff0000;
+ *(StgInt32 *)(adj_code+0x20) = 0x00000014;
+
+ *(StgInt64 *)(adj_code+0x28) = (StgInt64)obscure_ccall_ret_code;
+ *(StgInt64 *)(adj_code+0x30) = (StgInt64)hptr;
+ *(StgInt64 *)(adj_code+0x38) = (StgInt64)wptr;
+ }
+ }
+#endif /* defined(mingw32_HOST_OS) */
+
+ default:
+ barf("createAdjustor: Unsupported calling convention");
+ break;
+ }
+
+ return code;
+}
+
+void freeHaskellFunctionPtr(void* ptr)
+{
+ if ( *(StgWord16 *)ptr == 0x894d ) {
+ freeStablePtr(*(StgStablePtr*)((StgWord8*)ptr+
+#if defined(mingw32_HOST_OS)
+ 0x28
+#else
+ 0x20
+#endif
+ ));
+#if !defined(mingw32_HOST_OS)
+ } else if ( *(StgWord16 *)ptr == 0x5141 ) {
+ freeStablePtr(*(StgStablePtr*)((StgWord8*)ptr+0x30));
+#endif
+#if defined(mingw32_HOST_OS)
+ } else if ( *(StgWord16 *)ptr == 0x8348 ) {
+ freeStablePtr(*(StgStablePtr*)((StgWord8*)ptr+0x48));
+#endif
+ } else {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+
+ // Can't write to this memory, it is only executable:
+ // *((unsigned char*)ptr) = '\0';
+
+ freeExec(ptr);
+}
diff --git a/rts/adjustor/NativeIA64.c b/rts/adjustor/NativeIA64.c
new file mode 100644
index 0000000000..9fd1991c0a
--- /dev/null
+++ b/rts/adjustor/NativeIA64.c
@@ -0,0 +1,165 @@
+/* -----------------------------------------------------------------------------
+ * IA64 architecture adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+
+/* Layout of a function descriptor */
+typedef struct _IA64FunDesc {
+ StgWord64 ip;
+ StgWord64 gp;
+} IA64FunDesc;
+
+static void *
+stgAllocStable(size_t size_in_bytes, StgStablePtr *stable)
+{
+ StgArrBytes* arr;
+ uint32_t data_size_in_words, total_size_in_words;
+
+ /* round up to a whole number of words */
+ data_size_in_words = ROUNDUP_BYTES_TO_WDS(size_in_bytes);
+ total_size_in_words = sizeofW(StgArrBytes) + data_size_in_words;
+
+ /* allocate and fill it in */
+ arr = (StgArrBytes *)allocate(total_size_in_words);
+ SET_ARR_HDR(arr, &stg_ARR_WORDS_info, CCCS, size_in_bytes);
+
+ /* obtain a stable ptr */
+ *stable = getStablePtr((StgPtr)arr);
+
+ /* and return a ptr to the goods inside the array */
+ return(&(arr->payload));
+}
+
+void*
+createAdjustor(int cconv, StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString
+#if !defined(powerpc_HOST_ARCH) && !defined(powerpc64_HOST_ARCH) && !defined(x86_64_HOST_ARCH)
+ STG_UNUSED
+#endif
+ )
+{
+ void *adjustor = NULL;
+ void *code = NULL;
+
+ switch (cconv)
+ {
+ case 1: /* _ccall */
+/*
+ Up to 8 inputs are passed in registers. We flush the last two inputs to
+ the stack, initially into the 16-byte scratch region left by the caller.
+ We then shuffle the others along by 4 (taking 2 registers for ourselves
+ to save return address and previous function state - we need to come back
+ here on the way out to restore the stack, so this is a real function
+ rather than just a trampoline).
+
+ The function descriptor we create contains the gp of the target function
+ so gp is already loaded correctly.
+
+ [MLX] alloc r16=ar.pfs,10,2,0
+ movl r17=wptr
+ [MII] st8.spill [r12]=r38,8 // spill in6 (out4)
+ mov r41=r37 // out7 = in5 (out3)
+ mov r40=r36;; // out6 = in4 (out2)
+ [MII] st8.spill [r12]=r39 // spill in7 (out5)
+ mov.sptk b6=r17,50
+ mov r38=r34;; // out4 = in2 (out0)
+ [MII] mov r39=r35 // out5 = in3 (out1)
+ mov r37=r33 // out3 = in1 (loc1)
+ mov r36=r32 // out2 = in0 (loc0)
+ [MLX] adds r12=-24,r12 // update sp
+ movl r34=hptr;; // out0 = hptr
+ [MIB] mov r33=r16 // loc1 = ar.pfs
+ mov r32=b0 // loc0 = retaddr
+ br.call.sptk.many b0=b6;;
+
+ [MII] adds r12=-16,r12
+ mov b0=r32
+ mov.i ar.pfs=r33
+ [MFB] nop.m 0x0
+ nop.f 0x0
+ br.ret.sptk.many b0;;
+*/
+
+/* These macros distribute a long constant into the two words of an MLX bundle */
+#define BITS(val,start,count) (((val) >> (start)) & ((1 << (count))-1))
+#define MOVL_LOWORD(val) (BITS(val,22,18) << 46)
+#define MOVL_HIWORD(val) ( (BITS(val,0,7) << 36) \
+ | (BITS(val,7,9) << 50) \
+ | (BITS(val,16,5) << 45) \
+ | (BITS(val,21,1) << 44) \
+ | (BITS(val,40,23)) \
+ | (BITS(val,63,1) << 59))
+
+ {
+ StgStablePtr stable;
+ IA64FunDesc *wdesc = (IA64FunDesc *)wptr;
+ StgWord64 wcode = wdesc->ip;
+ IA64FunDesc *fdesc;
+ StgWord64 *code;
+
+ /* we allocate on the Haskell heap since malloc'd memory isn't
+ * executable - argh */
+ /* Allocated memory is word-aligned (8 bytes) but functions on ia64
+ * must be aligned to 16 bytes. We allocate an extra 8 bytes of
+ * wiggle room so that we can put the code on a 16 byte boundary. */
+ adjustor = stgAllocStable(sizeof(IA64FunDesc)+18*8+8, &stable);
+
+ fdesc = (IA64FunDesc *)adjustor;
+ code = (StgWord64 *)(fdesc + 1);
+ /* add 8 bytes to code if needed to align to a 16-byte boundary */
+ if ((StgWord64)code & 15) code++;
+ fdesc->ip = (StgWord64)code;
+ fdesc->gp = wdesc->gp;
+
+ code[0] = 0x0000058004288004 | MOVL_LOWORD(wcode);
+ code[1] = 0x6000000220000000 | MOVL_HIWORD(wcode);
+ code[2] = 0x029015d818984001;
+ code[3] = 0x8401200500420094;
+ code[4] = 0x886011d8189c0001;
+ code[5] = 0x84011004c00380c0;
+ code[6] = 0x0250210046013800;
+ code[7] = 0x8401000480420084;
+ code[8] = 0x0000233f19a06005 | MOVL_LOWORD((StgWord64)hptr);
+ code[9] = 0x6000000440000000 | MOVL_HIWORD((StgWord64)hptr);
+ code[10] = 0x0200210020010811;
+ code[11] = 0x1080006800006200;
+ code[12] = 0x0000210018406000;
+ code[13] = 0x00aa021000038005;
+ code[14] = 0x000000010000001d;
+ code[15] = 0x0084000880000200;
+
+ /* save stable pointers in convenient form */
+ code[16] = (StgWord64)hptr;
+ code[17] = (StgWord64)stable;
+ }
+#else
+ barf("adjustor creation not supported on this platform");
+#endif
+ break;
+
+ default:
+ barf("createAdjustor: Unsupported calling convention");
+ }
+
+ return code;
+}
+
+void
+freeHaskellFunctionPtr(void* ptr)
+{
+ IA64FunDesc *fdesc = (IA64FunDesc *)ptr;
+ StgWord64 *code = (StgWord64 *)(fdesc+1);
+
+ if (fdesc->ip != (StgWord64)code) {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+ freeStablePtr((StgStablePtr)code[16]);
+ freeStablePtr((StgStablePtr)code[17]);
+}
diff --git a/rts/adjustor/NativePowerPC.c b/rts/adjustor/NativePowerPC.c
new file mode 100644
index 0000000000..e5612704e9
--- /dev/null
+++ b/rts/adjustor/NativePowerPC.c
@@ -0,0 +1,408 @@
+/* -----------------------------------------------------------------------------
+ * PowerPC architecture adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+#include "Adjustor.h"
+
+/* Adjustor logic for PowerPC and PowerPC64 */
+
+#if defined(linux_HOST_OS)
+#include <string.h>
+#endif
+
+// from AdjustorAsm.s
+// not declared as a function so that AIX-style
+// fundescs can never get in the way.
+extern void *adjustorCode;
+
+#if defined(linux_HOST_OS)
+__asm__("obscure_ccall_ret_code:\n\t"
+ "lwz 1,0(1)\n\t"
+ "lwz 0,4(1)\n\t"
+ "mtlr 0\n\t"
+ "blr");
+extern void obscure_ccall_ret_code(void);
+#endif /* defined(linux_HOST_OS) */
+
+#if defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH)
+#if !(defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS))
+
+/* !!! !!! WARNING: !!! !!!
+ * This structure is accessed from AdjustorAsm.s
+ * Any changes here have to be mirrored in the offsets there.
+ */
+
+typedef struct AdjustorStub {
+ /* fundesc-based ABIs */
+#define FUNDESCS
+ StgFunPtr code;
+ struct AdjustorStub
+ *toc;
+ void *env;
+ StgStablePtr hptr;
+ StgFunPtr wptr;
+ StgInt negative_framesize;
+ StgInt extrawords_plus_one;
+} AdjustorStub;
+
+#endif /* !(defined(powerpc_HOST_ARCH) && defined(linux_HOST_OS)) */
+#endif /* defined(powerpc_HOST_ARCH) || defined(powerpc64_HOST_ARCH) */
+
+void*
+createAdjustor(int cconv, StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString
+ )
+{
+ void *adjustor = NULL;
+ void *code = NULL;
+
+ switch (cconv)
+ {
+ case 1: /* _ccall */
+#if defined(linux_HOST_OS)
+
+#define OP_LO(op,lo) ((((unsigned)(op)) << 16) | (((unsigned)(lo)) & 0xFFFF))
+#define OP_HI(op,hi) ((((unsigned)(op)) << 16) | (((unsigned)(hi)) >> 16))
+ {
+ /* The PowerPC Linux (32-bit) calling convention is annoyingly complex.
+ We need to calculate all the details of the stack frame layout,
+ taking into account the types of all the arguments, and then
+ generate code on the fly. */
+
+ int src_gpr = 3, dst_gpr = 5;
+ int fpr = 3;
+ int src_offset = 0, dst_offset = 0;
+ int n = strlen(typeString),i;
+ int src_locs[n], dst_locs[n];
+ int frameSize;
+ unsigned *code;
+
+ /* Step 1:
+ Calculate where the arguments should go.
+ src_locs[] will contain the locations of the arguments in the
+ original stack frame passed to the adjustor.
+ dst_locs[] will contain the locations of the arguments after the
+ adjustor runs, on entry to the wrapper proc pointed to by wptr.
+
+ This algorithm is based on the one described on page 3-19 of the
+ System V ABI PowerPC Processor Supplement.
+ */
+ for(i=0;typeString[i];i++)
+ {
+ char t = typeString[i];
+ if((t == 'f' || t == 'd') && fpr <= 8)
+ src_locs[i] = dst_locs[i] = -32-(fpr++);
+ else
+ {
+ if((t == 'l' || t == 'L') && src_gpr <= 9)
+ {
+ if((src_gpr & 1) == 0)
+ src_gpr++;
+ src_locs[i] = -src_gpr;
+ src_gpr += 2;
+ }
+ else if((t == 'w' || t == 'W') && src_gpr <= 10)
+ {
+ src_locs[i] = -(src_gpr++);
+ }
+ else
+ {
+ if(t == 'l' || t == 'L' || t == 'd')
+ {
+ if(src_offset % 8)
+ src_offset += 4;
+ }
+ src_locs[i] = src_offset;
+ src_offset += (t == 'l' || t == 'L' || t == 'd') ? 8 : 4;
+ }
+
+ if((t == 'l' || t == 'L') && dst_gpr <= 9)
+ {
+ if((dst_gpr & 1) == 0)
+ dst_gpr++;
+ dst_locs[i] = -dst_gpr;
+ dst_gpr += 2;
+ }
+ else if((t == 'w' || t == 'W') && dst_gpr <= 10)
+ {
+ dst_locs[i] = -(dst_gpr++);
+ }
+ else
+ {
+ if(t == 'l' || t == 'L' || t == 'd')
+ {
+ if(dst_offset % 8)
+ dst_offset += 4;
+ }
+ dst_locs[i] = dst_offset;
+ dst_offset += (t == 'l' || t == 'L' || t == 'd') ? 8 : 4;
+ }
+ }
+ }
+
+ frameSize = dst_offset + 8;
+ frameSize = (frameSize+15) & ~0xF;
+
+ /* Step 2:
+ Build the adjustor.
+ */
+ // allocate space for at most 4 insns per parameter
+ // plus 14 more instructions.
+ adjustor = allocateExec(4 * (4*n + 14),&code);
+ code = (unsigned*)adjustor;
+
+ *code++ = 0x48000008; // b *+8
+ // * Put the hptr in a place where freeHaskellFunctionPtr
+ // can get at it.
+ *code++ = (unsigned) hptr;
+
+ // * save the link register
+ *code++ = 0x7c0802a6; // mflr r0;
+ *code++ = 0x90010004; // stw r0, 4(r1);
+ // * and build a new stack frame
+ *code++ = OP_LO(0x9421, -frameSize); // stwu r1, -frameSize(r1)
+
+ // * now generate instructions to copy arguments
+ // from the old stack frame into the new stack frame.
+ for(i=n-1;i>=0;i--)
+ {
+ if(src_locs[i] < -32)
+ ASSERT(dst_locs[i] == src_locs[i]);
+ else if(src_locs[i] < 0)
+ {
+ // source in GPR.
+ ASSERT(typeString[i] != 'f' && typeString[i] != 'd');
+ if(dst_locs[i] < 0)
+ {
+ ASSERT(dst_locs[i] > -32);
+ // dst is in GPR, too.
+
+ if(typeString[i] == 'l' || typeString[i] == 'L')
+ {
+ // mr dst+1, src+1
+ *code++ = 0x7c000378
+ | ((-dst_locs[i]+1) << 16)
+ | ((-src_locs[i]+1) << 11)
+ | ((-src_locs[i]+1) << 21);
+ }
+ // mr dst, src
+ *code++ = 0x7c000378
+ | ((-dst_locs[i]) << 16)
+ | ((-src_locs[i]) << 11)
+ | ((-src_locs[i]) << 21);
+ }
+ else
+ {
+ if(typeString[i] == 'l' || typeString[i] == 'L')
+ {
+ // stw src+1, dst_offset+4(r1)
+ *code++ = 0x90010000
+ | ((-src_locs[i]+1) << 21)
+ | (dst_locs[i] + 4);
+ }
+
+ // stw src, dst_offset(r1)
+ *code++ = 0x90010000
+ | ((-src_locs[i]) << 21)
+ | (dst_locs[i] + 8);
+ }
+ }
+ else
+ {
+ ASSERT(dst_locs[i] >= 0);
+ ASSERT(typeString[i] != 'f' && typeString[i] != 'd');
+
+ if(typeString[i] == 'l' || typeString[i] == 'L')
+ {
+ // lwz r0, src_offset(r1)
+ *code++ = 0x80010000
+ | (src_locs[i] + frameSize + 8 + 4);
+ // stw r0, dst_offset(r1)
+ *code++ = 0x90010000
+ | (dst_locs[i] + 8 + 4);
+ }
+ // lwz r0, src_offset(r1)
+ *code++ = 0x80010000
+ | (src_locs[i] + frameSize + 8);
+ // stw r0, dst_offset(r1)
+ *code++ = 0x90010000
+ | (dst_locs[i] + 8);
+ }
+ }
+
+ // * hptr will be the new first argument.
+ // lis r3, hi(hptr)
+ *code++ = OP_HI(0x3c60, hptr);
+ // ori r3,r3,lo(hptr)
+ *code++ = OP_LO(0x6063, hptr);
+
+ // * we need to return to a piece of code
+ // which will tear down the stack frame.
+ // lis r11,hi(obscure_ccall_ret_code)
+ *code++ = OP_HI(0x3d60, obscure_ccall_ret_code);
+ // ori r11,r11,lo(obscure_ccall_ret_code)
+ *code++ = OP_LO(0x616b, obscure_ccall_ret_code);
+ // mtlr r11
+ *code++ = 0x7d6803a6;
+
+ // * jump to wptr
+ // lis r11,hi(wptr)
+ *code++ = OP_HI(0x3d60, wptr);
+ // ori r11,r11,lo(wptr)
+ *code++ = OP_LO(0x616b, wptr);
+ // mtctr r11
+ *code++ = 0x7d6903a6;
+ // bctr
+ *code++ = 0x4e800420;
+
+ // Flush the Instruction cache:
+ {
+ unsigned *p = adjustor;
+ while(p < code)
+ {
+ __asm__ volatile ("dcbf 0,%0\n\tsync\n\ticbi 0,%0"
+ : : "r" (p));
+ p++;
+ }
+ __asm__ volatile ("sync\n\tisync");
+ }
+ }
+
+#else
+
+#define OP_LO(op,lo) ((((unsigned)(op)) << 16) | (((unsigned)(lo)) & 0xFFFF))
+#define OP_HI(op,hi) ((((unsigned)(op)) << 16) | (((unsigned)(hi)) >> 16))
+ {
+ /* The following code applies to all PowerPC and PowerPC64 platforms
+ whose stack layout is based on the AIX ABI.
+
+ Besides (obviously) AIX, this includes
+ Mac OS 9 and BeOS/PPC and Mac OS X PPC (may they rest in peace),
+ which use the 32-bit AIX ABI
+ powerpc64-linux,
+ which uses the 64-bit AIX ABI.
+
+ The actual stack-frame shuffling is implemented out-of-line
+ in the function adjustorCode, in AdjustorAsm.S.
+ Here, we set up an AdjustorStub structure, which
+ is a function descriptor with a pointer to the AdjustorStub
+ struct in the position of the TOC that is loaded
+ into register r2.
+
+ One nice thing about this is that there is _no_ code generated at
+ runtime on the platforms that have function descriptors.
+ */
+ AdjustorStub *adjustorStub;
+ int sz = 0, extra_sz, total_sz;
+
+#if defined(FUNDESCS)
+ adjustorStub = stgMallocBytes(sizeof(AdjustorStub), "createAdjustor");
+#else
+ adjustorStub = allocateExec(sizeof(AdjustorStub),&code);
+#endif /* defined(FUNDESCS) */
+ adjustor = adjustorStub;
+
+ adjustorStub->code = (void*) &adjustorCode;
+
+#if defined(FUNDESCS)
+ // function descriptors are a cool idea.
+ // We don't need to generate any code at runtime.
+ adjustorStub->toc = adjustorStub;
+#else
+
+ // no function descriptors :-(
+ // We need to do things "by hand".
+#if defined(powerpc_HOST_ARCH)
+ // lis r2, hi(adjustorStub)
+ adjustorStub->lis = OP_HI(0x3c40, adjustorStub);
+ // ori r2, r2, lo(adjustorStub)
+ adjustorStub->ori = OP_LO(0x6042, adjustorStub);
+ // lwz r0, code(r2)
+ adjustorStub->lwz = OP_LO(0x8002, (char*)(&adjustorStub->code)
+ - (char*)adjustorStub);
+ // mtctr r0
+ adjustorStub->mtctr = 0x7c0903a6;
+ // bctr
+ adjustorStub->bctr = 0x4e800420;
+#else
+ barf("adjustor creation not supported on this platform");
+#endif /* defined(powerpc_HOST_ARCH) */
+
+ // Flush the Instruction cache:
+ {
+ int n = sizeof(AdjustorStub)/sizeof(unsigned);
+ unsigned *p = (unsigned*)adjustor;
+ while(n--)
+ {
+ __asm__ volatile ("dcbf 0,%0\n\tsync\n\ticbi 0,%0"
+ : : "r" (p));
+ p++;
+ }
+ __asm__ volatile ("sync\n\tisync");
+ }
+#endif /* defined(FUNDESCS) */
+
+ // Calculate the size of the stack frame, in words.
+ sz = totalArgumentSize(typeString);
+
+ // The first eight words of the parameter area
+ // are just "backing store" for the parameters passed in
+ // the GPRs. extra_sz is the number of words beyond those first
+ // 8 words.
+ extra_sz = sz - 8;
+ if(extra_sz < 0)
+ extra_sz = 0;
+
+ // Calculate the total size of the stack frame.
+ total_sz = (6 /* linkage area */
+ + 8 /* minimum parameter area */
+ + 2 /* two extra arguments */
+ + extra_sz)*sizeof(StgWord);
+
+ // align to 16 bytes.
+ // AIX only requires 8 bytes, but who cares?
+ total_sz = (total_sz+15) & ~0xF;
+
+ // Fill in the information that adjustorCode in AdjustorAsm.S
+ // will use to create a new stack frame with the additional args.
+ adjustorStub->hptr = hptr;
+ adjustorStub->wptr = wptr;
+ adjustorStub->negative_framesize = -total_sz;
+ adjustorStub->extrawords_plus_one = extra_sz + 1;
+ }
+
+ default:
+ barf("createAdjustor: Unsupported calling convention");
+ }
+
+ return code;
+}
+
+void
+freeHaskellFunctionPtr(void* ptr)
+{
+#if defined(linux_HOST_OS)
+ if ( *(StgWord*)ptr != 0x48000008 ) {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+ freeStablePtr(((StgStablePtr*)ptr)[1]);
+#else
+ if ( ((AdjustorStub*)ptr)->code != (StgFunPtr) &adjustorCode ) {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+ freeStablePtr(((AdjustorStub*)ptr)->hptr);
+#endif
+
+ // Can't write to this memory, it is only executable:
+ // *((unsigned char*)ptr) = '\0';
+
+ freeExec(ptr);
+}
diff --git a/rts/adjustor/NativeSparc.c b/rts/adjustor/NativeSparc.c
new file mode 100644
index 0000000000..a315cc2d94
--- /dev/null
+++ b/rts/adjustor/NativeSparc.c
@@ -0,0 +1,110 @@
+/* -----------------------------------------------------------------------------
+ * SPARC architecture adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+
+void*
+createAdjustor(int cconv, StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString STG_UNUSED
+ )
+{
+ void *adjustor = NULL;
+ void *code = NULL;
+
+ switch (cconv)
+ {
+ case 1: /* _ccall */
+ /* Magic constant computed by inspecting the code length of the following
+ assembly language snippet (offset and machine code prefixed):
+
+ <00>: 9C23A008 sub %sp, 8, %sp ! make room for %o4/%o5 in caller's frame
+ <04>: DA23A060 st %o5, [%sp + 96] ! shift registers by 2 positions
+ <08>: D823A05C st %o4, [%sp + 92]
+ <0C>: 9A10000B mov %o3, %o5
+ <10>: 9810000A mov %o2, %o4
+ <14>: 96100009 mov %o1, %o3
+ <18>: 94100008 mov %o0, %o2
+ <1C>: 13000000 sethi %hi(wptr), %o1 ! load up wptr (1 of 2)
+ <20>: 11000000 sethi %hi(hptr), %o0 ! load up hptr (1 of 2)
+ <24>: 81C26000 jmp %o1 + %lo(wptr) ! jump to wptr (load 2 of 2)
+ <28>: 90122000 or %o0, %lo(hptr), %o0 ! load up hptr (2 of 2, delay slot)
+ <2C> 00000000 ! place for getting hptr back easily
+
+ ccall'ing on SPARC is easy, because we are quite lucky to push a
+ multiple of 8 bytes (1 word hptr + 1 word dummy arg) in front of the
+ existing arguments (note that %sp must stay double-word aligned at
+ all times, see ABI spec at http://www.sparc.org/standards/psABI3rd.pdf).
+ To do this, we extend the *caller's* stack frame by 2 words and shift
+ the output registers used for argument passing (%o0 - %o5, we are a *leaf*
+ procedure because of the tail-jump) by 2 positions. This makes room in
+ %o0 and %o1 for the additional arguments, namely hptr and a dummy (used
+ for destination addr of jump on SPARC, return address on x86, ...). This
+ shouldn't cause any problems for a C-like caller: alloca is implemented
+ similarly, and local variables should be accessed via %fp, not %sp. In a
+ nutshell: This should work! (Famous last words! :-)
+ */
+ adjustor = allocateExec(4*(11+1),&code);
+ {
+ unsigned long *const adj_code = (unsigned long *)adjustor;
+
+ adj_code[ 0] = 0x9C23A008UL; /* sub %sp, 8, %sp */
+ adj_code[ 1] = 0xDA23A060UL; /* st %o5, [%sp + 96] */
+ adj_code[ 2] = 0xD823A05CUL; /* st %o4, [%sp + 92] */
+ adj_code[ 3] = 0x9A10000BUL; /* mov %o3, %o5 */
+ adj_code[ 4] = 0x9810000AUL; /* mov %o2, %o4 */
+ adj_code[ 5] = 0x96100009UL; /* mov %o1, %o3 */
+ adj_code[ 6] = 0x94100008UL; /* mov %o0, %o2 */
+ adj_code[ 7] = 0x13000000UL; /* sethi %hi(wptr), %o1 */
+ adj_code[ 7] |= ((unsigned long)wptr) >> 10;
+ adj_code[ 8] = 0x11000000UL; /* sethi %hi(hptr), %o0 */
+ adj_code[ 8] |= ((unsigned long)hptr) >> 10;
+ adj_code[ 9] = 0x81C26000UL; /* jmp %o1 + %lo(wptr) */
+ adj_code[ 9] |= ((unsigned long)wptr) & 0x000003FFUL;
+ adj_code[10] = 0x90122000UL; /* or %o0, %lo(hptr), %o0 */
+ adj_code[10] |= ((unsigned long)hptr) & 0x000003FFUL;
+
+ adj_code[11] = (unsigned long)hptr;
+
+ /* flush cache */
+ asm("flush %0" : : "r" (adj_code ));
+ asm("flush %0" : : "r" (adj_code + 2));
+ asm("flush %0" : : "r" (adj_code + 4));
+ asm("flush %0" : : "r" (adj_code + 6));
+ asm("flush %0" : : "r" (adj_code + 10));
+
+ /* max. 5 instructions latency, and we need at >= 1 for returning */
+ asm("nop");
+ asm("nop");
+ asm("nop");
+ asm("nop");
+ }
+
+ default:
+ barf("createAdjustor: Unsupported calling convention");
+ }
+
+ return code;
+}
+
+void
+freeHaskellFunctionPtr(void* ptr)
+{
+ if ( *(unsigned long*)ptr != 0x9C23A008UL ) {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+
+ /* Free the stable pointer first..*/
+ freeStablePtr(*((StgStablePtr*)((unsigned long*)ptr + 11)));
+
+ // Can't write to this memory, it is only executable:
+ // *((unsigned char*)ptr) = '\0';
+
+ freeExec(ptr);
+}
diff --git a/rts/adjustor/Nativei386.c b/rts/adjustor/Nativei386.c
new file mode 100644
index 0000000000..976bc8381a
--- /dev/null
+++ b/rts/adjustor/Nativei386.c
@@ -0,0 +1,137 @@
+/* -----------------------------------------------------------------------------
+ * i386 architecture adjustor thunk logic.
+ * ---------------------------------------------------------------------------*/
+
+#include "PosixSource.h"
+#include "Rts.h"
+
+#include "RtsUtils.h"
+#include "StablePtr.h"
+#include "Adjustor.h"
+
+#if defined(_WIN32)
+#include <windows.h>
+#endif
+
+extern void adjustorCode(void);
+
+/* !!! !!! WARNING: !!! !!!
+ * This structure is accessed from AdjustorAsm.s
+ * Any changes here have to be mirrored in the offsets there.
+ */
+
+typedef struct AdjustorStub {
+ unsigned char call[8];
+ StgStablePtr hptr;
+ StgFunPtr wptr;
+ StgInt frame_size;
+ StgInt argument_size;
+} AdjustorStub;
+
+void*
+createAdjustor(int cconv, StgStablePtr hptr,
+ StgFunPtr wptr,
+ char *typeString STG_UNUSED
+ )
+{
+ void *adjustor = NULL;
+ void *code = NULL;
+
+ switch (cconv)
+ {
+ case 0: /* _stdcall */
+#if !defined(darwin_HOST_OS)
+ /* Magic constant computed by inspecting the code length of
+ the following assembly language snippet
+ (offset and machine code prefixed):
+
+ <0>: 58 popl %eax # temp. remove ret addr..
+ <1>: 68 fd fc fe fa pushl 0xfafefcfd # constant is large enough to
+ # hold a StgStablePtr
+ <6>: 50 pushl %eax # put back ret. addr
+ <7>: b8 fa ef ff 00 movl $0x00ffeffa, %eax # load up wptr
+ <c>: ff e0 jmp %eax # and jump to it.
+ # the callee cleans up the stack
+ */
+
+ {
+ unsigned char adj_code[14];
+ adj_code[0x00] = (unsigned char)0x58; /* popl %eax */
+
+ adj_code[0x01] = (unsigned char)0x68; /* pushl hptr (which is a dword immediate ) */
+ *((StgStablePtr*)(adj_code + 0x02)) = (StgStablePtr)hptr;
+
+ adj_code[0x06] = (unsigned char)0x50; /* pushl %eax */
+
+ adj_code[0x07] = (unsigned char)0xb8; /* movl $wptr, %eax */
+ *((StgFunPtr*)(adj_code + 0x08)) = (StgFunPtr)wptr;
+
+ adj_code[0x0c] = (unsigned char)0xff; /* jmp %eax */
+ adj_code[0x0d] = (unsigned char)0xe0;
+ adjustor = allocateExec(14, &adj_code);
+ }
+#endif /* !defined(darwin_HOST_OS) */
+
+ case 1: /* _ccall */
+ {
+ /*
+ Most of the trickiness here is due to the need to keep the
+ stack pointer 16-byte aligned (see #5250). That means we
+ can't just push another argument on the stack and call the
+ wrapper, we may have to shuffle the whole argument block.
+
+ We offload most of the work to AdjustorAsm.S.
+ */
+ AdjustorStub *adjustorStub = allocateExec(sizeof(AdjustorStub),&code);
+ adjustor = adjustorStub;
+
+ int sz = totalArgumentSize(typeString);
+
+ adjustorStub->call[0] = 0xe8;
+ *(long*)&adjustorStub->call[1] = ((char*)&adjustorCode) - ((char*)code + 5);
+ adjustorStub->hptr = hptr;
+ adjustorStub->wptr = wptr;
+
+ // The adjustor puts the following things on the stack:
+ // 1.) %ebp link
+ // 2.) padding and (a copy of) the arguments
+ // 3.) a dummy argument
+ // 4.) hptr
+ // 5.) return address (for returning to the adjustor)
+ // All these have to add up to a multiple of 16.
+
+ // first, include everything in frame_size
+ adjustorStub->frame_size = sz * 4 + 16;
+ // align to 16 bytes
+ adjustorStub->frame_size = (adjustorStub->frame_size + 15) & ~15;
+ // only count 2.) and 3.) as part of frame_size
+ adjustorStub->frame_size -= 12;
+ adjustorStub->argument_size = sz;
+ }
+
+ default:
+ barf("createAdjustor: Unsupported calling convention");
+ }
+
+ return code;
+}
+
+void
+freeHaskellFunctionPtr(void* ptr)
+{
+ if ( *(unsigned char*)ptr != 0xe8 &&
+ *(unsigned char*)ptr != 0x58 ) {
+ errorBelch("freeHaskellFunctionPtr: not for me, guv! %p\n", ptr);
+ return;
+ }
+ if (*(unsigned char*)ptr == 0xe8) { /* Aha, a ccall adjustor! */
+ freeStablePtr(((AdjustorStub*)ptr)->hptr);
+ } else {
+ freeStablePtr(*((StgStablePtr*)((unsigned char*)ptr + 0x02)));
+ }
+
+ // Can't write to this memory, it is only executable:
+ // *((unsigned char*)ptr) = '\0';
+
+ freeExec(ptr);
+}
diff --git a/rts/ghc.mk b/rts/ghc.mk
index 15854de5aa..d3e533d1f7 100644
--- a/rts/ghc.mk
+++ b/rts/ghc.mk
@@ -54,15 +54,31 @@ rts_CMM_SRCS := $(wildcard rts/*.cmm)
# Don't compile .S files when bootstrapping a new arch
ifneq "$(PORTING_HOST)" "YES"
-ifneq "$(findstring $(TargetArch_CPP), i386 powerpc powerpc64)" ""
-rts_S_SRCS += rts/AdjustorAsm.S
-endif
-# this matches substrings of powerpc64le, including "powerpc" and "powerpc64"
-ifneq "$(findstring $(TargetArch_CPP), powerpc64le s390x riscv64)" ""
+
# unregisterised builds use the mini interpreter
ifneq "$(GhcUnregisterised)" "YES"
rts_S_SRCS += rts/StgCRunAsm.S
endif
+
+# select adjustor implementation. This much match the logic in rts.cabal.in.
+ifeq "$(UseLibffiForAdjustors)" "YES"
+rts_C_SRCS += rts/adjustor/LibffiAdjustor.c
+else
+ifneq "$(findstring $(TargetArch_CPP), i386)" ""
+rts_S_SRCS += rts/AdjustorAsm.S
+rts_C_SRCS += rts/adjustor/Nativei386.c
+else
+ifneq "$(findstring $(TargetArch_CPP), x86_64)" ""
+rts_C_SRCS += rts/adjustor/NativeAmd64.c
+else
+ifneq "$(findstring $(TargetArch_CPP), powerpc64le powerpc)" ""
+rts_S_SRCS += rts/AdjustorAsm.S
+rts_C_SRCS += rts/adjustor/NativePowerPC.c
+else
+$(error Target architecture has no native adjustor implementation)
+endif
+endif
+endif
endif
endif
@@ -381,10 +397,6 @@ rts_CC_OPTS += -DNOSMP
rts_HC_OPTS += -optc-DNOSMP
endif
-ifeq "$(UseLibFFIForAdjustors)" "YES"
-rts_CC_OPTS += -DUSE_LIBFFI_FOR_ADJUSTORS
-endif
-
# We *want* type-checking of hand-written cmm.
rts_HC_OPTS += -dcmm-lint
diff --git a/rts/rts.cabal.in b/rts/rts.cabal.in
index 22846ecc61..7c0f40dcfa 100644
--- a/rts/rts.cabal.in
+++ b/rts/rts.cabal.in
@@ -12,6 +12,8 @@ flag libdl
default: @CabalHaveLibdl@
flag ffi
default: @CabalHaveLibffi@
+flag libffi-adjustors
+ default: @CabalLibffiAdjustors@
flag need-pthread
default: @CabalNeedLibpthread@
flag libbfd
@@ -417,10 +419,27 @@ library
-- AutoApply is generated
AutoApply.cmm
- if arch(i386) || arch(ppc) || arch(ppc64)
- asm-sources: AdjustorAsm.S
+ -- Adjustor stuff
+ if flag(libffi-adjustors)
+ c-sources: adjustor/LibffiAdjustor.c
+ else
+ -- Use GHC's native adjustors
+ if arch(i386)
+ asm-sources: AdjustorAsm.S
+ c-sources: adjustor/Nativei386.c
+ if arch(x86_64)
+ c-sources: adjustor/NativeAmd64.c
+ if arch(ppc) || arch(ppc64)
+ asm-sources: AdjustorAsm.S
+ c-sources: adjustor/NativePowerPC.c
+ if arch(ia64)
+ c-sources: adjustor/NativeIA64.c
+ if arch(sparc)
+ c-sources: adjustor/NativeSparc.c
+
+ -- Use assembler STG entrypoint on archictures where it is used
if arch(ppc) || arch(ppc64) || arch(s390x) || arch(riscv64)
- asm-sources: StgCRunAsm.S
+ asm-sources: StgCRunAsm.S
c-sources: Adjustor.c
Arena.c