/* ----------------------------------------------------------------------------
- * $Id: SMP.h,v 1.4 2002/02/04 20:10:47 sof Exp $
*
- * (c) The GHC Team, 1999
+ * (c) The GHC Team, 2005
*
* Macros for SMP support
*
#error Build options incompatible with SMP.
#endif
-/*
- * CMPXCHG - this instruction is the standard "test & set". We use it
- * for locking closures in the thunk and blackhole entry code. If the
- * closure is already locked, or has an unexpected info pointer
- * (because another thread is altering it in parallel), we just jump
- * to the new entry point.
+/*
+ * XCHG - the atomic exchange instruction. Used for locking closures
+ * during updates (see lockClosure() below) and the MVar primops.
+ *
+ * NB: the xchg instruction is implicitly locked, so we do not need
+ * a lock prefix here.
*/
-#if defined(i386_TARGET_ARCH) && defined(TABLES_NEXT_TO_CODE)
-#define CMPXCHG(p, cmp, new) \
- __asm__ __volatile__ ( \
- "lock ; cmpxchg %1, %0\n" \
- "\tje 1f\n" \
- "\tjmp *%%eax\n" \
- "\t1:\n" \
- : /* no outputs */ \
- : "m" (p), "r" (new), "r" (cmp) \
- )
+INLINE_HEADER StgWord
+xchg(StgPtr p, StgWord w)
+{
+ StgWord result;
+ result = w;
+ __asm__ __volatile__ (
+ "xchg %1,%0"
+ :"+r" (result), "+m" (*p)
+ : /* no input-only operands */
+ );
+ return result;
+}
/*
- * XCHG - the atomic exchange instruction. Used for locking closures
- * during updates (see LOCK_CLOSURE below) and the MVar primops.
+ * CMPXCHG - the single-word atomic compare-and-exchange instruction. Used
+ * in the STM implementation.
*/
-#define XCHG(reg, obj) \
- __asm__ __volatile__ ( \
- "xchgl %1,%0" \
- :"+r" (reg), "+m" (obj) \
- : /* no input-only operands */ \
- )
+INLINE_HEADER StgWord
+cas(StgVolatilePtr p, StgWord o, StgWord n)
+{
+ __asm__ __volatile__ (
+ "lock cmpxchg %3,%1"
+ :"=a"(o), "=m" (*(volatile unsigned int *)p)
+ :"0" (o), "r" (n));
+ return o;
+}
+/*
+ * Write barrier - ensure that all preceding writes have happened
+ * before all following writes.
+ *
+ * We need to tell both the compiler AND the CPU about the barrier.
+ * This is a brute force solution; better results might be obtained by
+ * using volatile type declarations to get fine-grained ordering
+ * control in C, and optionally a memory barrier instruction on CPUs
+ * that require it (not x86 or x86_64).
+ */
+INLINE_HEADER void
+wb(void) {
+#if i386_HOST_ARCH || x86_64_HOST_ARCH
+ __asm__ __volatile__ ("" : : : "memory");
#else
-#error SMP macros not defined for this architecture
+#error memory barriers unimplemented on this architecture
#endif
+}
/*
- * LOCK_CLOSURE locks the specified closure, busy waiting for any
- * existing locks to be cleared.
+ * Locking/unlocking closures
+ *
+ * This is used primarily in the implementation of MVars.
*/
-#define LOCK_CLOSURE(c) \
- ({ \
- const StgInfoTable *__info; \
- __info = &stg_WHITEHOLE_info; \
- do { \
- XCHG(__info,((StgClosure *)(c))->header.info); \
- } while (__info == &stg_WHITEHOLE_info); \
- __info; \
- })
+#define SPIN_COUNT 4000
-#define LOCK_THUNK(__info) \
- CMPXCHG(R1.cl->header.info, __info, &stg_WHITEHOLE_info);
+INLINE_HEADER StgInfoTable *
+lockClosure(StgClosure *p)
+{
+#if i386_HOST_ARCH || x86_64_HOST_ARCH
+ StgWord info;
+ do {
+ nat i = 0;
+ do {
+ info = xchg((P_)&p->header.info, (W_)&stg_WHITEHOLE_info);
+ if (info != (W_)&stg_WHITEHOLE_info) return (StgInfoTable *)info;
+ } while (++i < SPIN_COUNT);
+ yieldThread();
+ } while (1);
+#else
+ ACQUIRE_SM_LOCK
+#endif
+}
+
+INLINE_HEADER void
+unlockClosure(StgClosure *p, StgInfoTable *info)
+{
+#if i386_HOST_ARCH || x86_64_HOST_ARCH
+ // This is a strictly ordered write, so we need a wb():
+ wb();
+ p->header.info = info;
+#else
+ RELEASE_SM_LOCK;
+#endif
+}
#else /* !SMP */
-#define LOCK_CLOSURE(c) /* nothing */
-#define LOCK_THUNK(__info) /* nothing */
+#define wb() /* nothing */
+
+INLINE_HEADER StgWord
+xchg(StgPtr p, StgWord w)
+{
+ StgWord old = *p;
+ *p = w;
+ return old;
+}
-#endif /* SMP */
+#endif /* !SMP */
#endif /* SMP_H */