X-Git-Url: http://git.megacz.com/?p=ghc-hetmet.git;a=blobdiff_plain;f=includes%2FSMP.h;h=f574ec8100a97f9da7790b3d9425168b8fb6f8b4;hp=91ffc22383640cfd9172ff29fb62945717fa818c;hb=540311cfcf08b7184d7ab57245b18214bd4b9173;hpb=edad06244b0a542d6b5c9ce3f3cd72941c5be804 diff --git a/includes/SMP.h b/includes/SMP.h index 91ffc22..f574ec8 100644 --- a/includes/SMP.h +++ b/includes/SMP.h @@ -16,16 +16,54 @@ * Unregisterised builds are ok, but only 1 CPU supported. */ +#ifdef CMINUSMINUS + +#define unlockClosure(ptr,info) \ + prim %write_barrier() []; \ + StgHeader_info(ptr) = info; + +#else + #if defined(THREADED_RTS) #if defined(TICKY_TICKY) #error Build options incompatible with THREADED_RTS. #endif +/* ---------------------------------------------------------------------------- + Atomic operations + ------------------------------------------------------------------------- */ + +/* + * The atomic exchange operation: xchg(p,w) exchanges the value + * pointed to by p with the value w, returning the old value. + * + * Used for locking closures during updates (see lockClosure() below) + * and the MVar primops. + */ +INLINE_HEADER StgWord xchg(StgPtr p, StgWord w); + /* - * XCHG - the atomic exchange instruction. Used for locking closures - * during updates (see lockClosure() below) and the MVar primops. + * Compare-and-swap. Atomically does this: * + * cas(p,o,n) { + * r = *p; + * if (r == o) { *p = n }; + * return r; + * } + */ +INLINE_HEADER StgWord cas(StgVolatilePtr p, StgWord o, StgWord n); + +/* + * Prevents write operations from moving across this call in either + * direction. + */ +INLINE_HEADER void write_barrier(void); + +/* ---------------------------------------------------------------------------- + Implementations + ------------------------------------------------------------------------- */ +/* * NB: the xchg instruction is implicitly locked, so we do not need * a lock prefix here. */ @@ -45,9 +83,16 @@ xchg(StgPtr p, StgWord w) "1: lwarx %0, 0, %2\n" " stwcx. %1, 0, %2\n" " bne- 1b" - :"=r" (result) + :"=&r" (result) :"r" (w), "r" (p) ); +#elif sparc_HOST_ARCH + result = w; + __asm__ __volatile__ ( + "swap %1,%0" + : "+r" (result), "+m" (*p) + : /* no input-only operands */ + ); #elif !defined(WITHSMP) result = *p; *p = w; @@ -66,7 +111,7 @@ cas(StgVolatilePtr p, StgWord o, StgWord n) { #if i386_HOST_ARCH || x86_64_HOST_ARCH __asm__ __volatile__ ( - "lock/cmpxchg %3,%1" + "lock\ncmpxchg %3,%1" :"=a"(o), "=m" (*(volatile unsigned int *)p) :"0" (o), "r" (n)); return o; @@ -84,6 +129,14 @@ cas(StgVolatilePtr p, StgWord o, StgWord n) :"cc", "memory" ); return result; +#elif sparc_HOST_ARCH + __asm__ __volatile__ ( + "cas [%1], %2, %0" + : "+r" (n) + : "r" (p), "r" (o) + : "memory" + ); + return n; #elif !defined(WITHSMP) StgWord result; result = *p; @@ -112,6 +165,9 @@ write_barrier(void) { __asm__ __volatile__ ("" : : : "memory"); #elif powerpc_HOST_ARCH __asm__ __volatile__ ("lwsync" : : : "memory"); +#elif sparc_HOST_ARCH + /* Sparc in TSO mode does not require write/write barriers. */ + __asm__ __volatile__ ("" : : : "memory"); #elif !defined(WITHSMP) return; #else @@ -119,14 +175,23 @@ write_barrier(void) { #endif } -/* +/* ----------------------------------------------------------------------------- * Locking/unlocking closures * * This is used primarily in the implementation of MVars. - */ + * -------------------------------------------------------------------------- */ + #define SPIN_COUNT 4000 -INLINE_HEADER StgInfoTable * +#ifdef KEEP_LOCKCLOSURE +// We want a callable copy of lockClosure() so that we can refer to it +// from .cmm files compiled using the native codegen. +extern StgInfoTable *lockClosure(StgClosure *p); +INLINE_ME +#else +INLINE_HEADER +#endif +StgInfoTable * lockClosure(StgClosure *p) { StgWord info; @@ -143,11 +208,100 @@ lockClosure(StgClosure *p) INLINE_HEADER void unlockClosure(StgClosure *p, StgInfoTable *info) { - // This is a strictly ordered write, so we need a wb(): + // This is a strictly ordered write, so we need a write_barrier(): write_barrier(); p->header.info = info; } +/* ----------------------------------------------------------------------------- + * Spin locks + * + * These are simple spin-only locks as opposed to Mutexes which + * probably spin for a while before blocking in the kernel. We use + * these when we are sure that all our threads are actively running on + * a CPU, eg. in the GC. + * + * TODO: measure whether we really need these, or whether Mutexes + * would do (and be a bit safer if a CPU becomes loaded). + * -------------------------------------------------------------------------- */ + +#if defined(DEBUG) +typedef struct StgSync_ +{ + StgWord32 lock; + StgWord64 spin; // DEBUG version counts how much it spins +} StgSync; +#else +typedef StgWord StgSync; +#endif + +typedef lnat StgSyncCount; + + +#if defined(DEBUG) + +// Debug versions of spin locks maintain a spin count + +// How to use: +// To use the debug veriosn of the spin locks, a debug version of the program +// can be run under a deugger with a break point on stat_exit. At exit time +// of the program one can examine the state the spin count counts of various +// spin locks to check for contention. + +// acquire spin lock +INLINE_HEADER void ACQUIRE_SPIN_LOCK(StgSync * p) +{ + StgWord32 r = 0; + do { + p->spin++; + r = cas((StgVolatilePtr)&(p->lock), 1, 0); + } while(r == 0); + p->spin--; +} + +// release spin lock +INLINE_HEADER void RELEASE_SPIN_LOCK(StgSync * p) +{ + write_barrier(); + p->lock = 1; +} + +// initialise spin lock +INLINE_HEADER void initSpinLock(StgSync * p) +{ + write_barrier(); + p->lock = 1; + p->spin = 0; +} + +#else + +// acquire spin lock +INLINE_HEADER void ACQUIRE_SPIN_LOCK(StgSync * p) +{ + StgWord32 r = 0; + do { + r = cas((StgVolatilePtr)p, 1, 0); + } while(r == 0); +} + +// release spin lock +INLINE_HEADER void RELEASE_SPIN_LOCK(StgSync * p) +{ + write_barrier(); + (*p) = 1; +} + +// init spin lock +INLINE_HEADER void initSpinLock(StgSync * p) +{ + write_barrier(); + (*p) = 1; +} + +#endif /* DEBUG */ + +/* ---------------------------------------------------------------------- */ #else /* !THREADED_RTS */ #define write_barrier() /* nothing */ @@ -168,6 +322,13 @@ INLINE_HEADER void unlockClosure(StgClosure *p STG_UNUSED, StgInfoTable *info STG_UNUSED) { /* nothing */ } +// Using macros here means we don't have to ensure the argument is in scope +#define ACQUIRE_SPIN_LOCK(p) /* nothing */ +#define RELEASE_SPIN_LOCK(p) /* nothing */ + +INLINE_HEADER void initSpinLock(void * p STG_UNUSED) +{ /* nothing */ } + #endif /* !THREADED_RTS */ // Handy specialised versions of lockClosure()/unlockClosure() @@ -178,3 +339,5 @@ INLINE_HEADER void unlockTSO(StgTSO *tso) { unlockClosure((StgClosure*)tso, (StgInfoTable*)&stg_TSO_info); } #endif /* SMP_H */ + +#endif /* CMINUSMINUS */