X-Git-Url: http://git.megacz.com/?a=blobdiff_plain;f=ghc%2Fincludes%2FSMP.h;h=d16d4ba64ec62d8719afe838b3b09931c742c76d;hb=6ff106616c1c98542df0a1763f0f9f655b2c8dd5;hp=86930f9b4ec055e765b19b9ee47040e2a58607ef;hpb=bf8219815876579f80b4e30188b982bf3e673199;p=ghc-hetmet.git diff --git a/ghc/includes/SMP.h b/ghc/includes/SMP.h index 86930f9..d16d4ba 100644 --- a/ghc/includes/SMP.h +++ b/ghc/includes/SMP.h @@ -2,30 +2,32 @@ * * (c) The GHC Team, 2005 * - * Macros for SMP support + * Macros for THREADED_RTS support * * -------------------------------------------------------------------------- */ #ifndef SMP_H #define SMP_H -/* SMP is currently not compatible with the following options: +/* THREADED_RTS is currently not compatible with the following options: * - * INTERPRETER - * PROFILING + * PROFILING (but only 1 CPU supported) * TICKY_TICKY - * and unregisterised builds. + * Unregisterised builds are ok, but only 1 CPU supported. */ -#if defined(SMP) +#if defined(THREADED_RTS) -#if defined(PROFILING) || defined(TICKY_TICKY) -#error Build options incompatible with SMP. +#if defined(TICKY_TICKY) +#error Build options incompatible with THREADED_RTS. #endif /* * XCHG - the atomic exchange instruction. Used for locking closures - * during updates (see LOCK_CLOSURE below) and the MVar primops. + * during updates (see lockClosure() below) and the MVar primops. + * + * NB: the xchg instruction is implicitly locked, so we do not need + * a lock prefix here. */ INLINE_HEADER StgWord xchg(StgPtr p, StgWord w) @@ -33,28 +35,95 @@ xchg(StgPtr p, StgWord w) StgWord result; result = w; __asm__ __volatile__ ( - "xchgl %1,%0" + "xchg %1,%0" :"+r" (result), "+m" (*p) : /* no input-only operands */ ); return result; } +/* + * CMPXCHG - the single-word atomic compare-and-exchange instruction. Used + * in the STM implementation. + */ +INLINE_HEADER StgWord +cas(StgVolatilePtr p, StgWord o, StgWord n) +{ + __asm__ __volatile__ ( + "lock cmpxchg %3,%1" + :"=a"(o), "=m" (*(volatile unsigned int *)p) + :"0" (o), "r" (n)); + return o; +} + +/* + * Write barrier - ensure that all preceding writes have happened + * before all following writes. + * + * We need to tell both the compiler AND the CPU about the barrier. + * This is a brute force solution; better results might be obtained by + * using volatile type declarations to get fine-grained ordering + * control in C, and optionally a memory barrier instruction on CPUs + * that require it (not x86 or x86_64). + */ +INLINE_HEADER void +wb(void) { +#if i386_HOST_ARCH || x86_64_HOST_ARCH + __asm__ __volatile__ ("" : : : "memory"); +#else +#error memory barriers unimplemented on this architecture +#endif +} + +/* + * Locking/unlocking closures + * + * This is used primarily in the implementation of MVars. + */ +#define SPIN_COUNT 4000 + INLINE_HEADER StgInfoTable * lockClosure(StgClosure *p) { +#if i386_HOST_ARCH || x86_64_HOST_ARCH StgWord info; -#if 0 do { - info = xchg((P_)&p->header.info, (W_)&stg_WHITEHOLE_info); - if (info != (W_)&stg_WHITEHOLE_info) return (StgInfoTable *)info; + nat i = 0; + do { + info = xchg((P_)&p->header.info, (W_)&stg_WHITEHOLE_info); + if (info != (W_)&stg_WHITEHOLE_info) return (StgInfoTable *)info; + } while (++i < SPIN_COUNT); yieldThread(); } while (1); #else - info = p->header.info; + ACQUIRE_SM_LOCK +#endif +} + +INLINE_HEADER void +unlockClosure(StgClosure *p, StgInfoTable *info) +{ +#if i386_HOST_ARCH || x86_64_HOST_ARCH + // This is a strictly ordered write, so we need a wb(): + wb(); + p->header.info = info; +#else + RELEASE_SM_LOCK; #endif } -#endif /* SMP */ +#else /* !THREADED_RTS */ + +#define wb() /* nothing */ + +INLINE_HEADER StgWord +xchg(StgPtr p, StgWord w) +{ + StgWord old = *p; + *p = w; + return old; +} + +#endif /* !THREADED_RTS */ #endif /* SMP_H */