#include "OSThreads.h"
#include "Capability.h"
+#include "EventLog.h"
/* initScheduler(), exitScheduler()
* Called from STG : no
* Locks assumed : none
*/
void initScheduler (void);
-void exitScheduler (void);
+void exitScheduler (rtsBool wait_foreign);
+void freeScheduler (void);
// Place a new thread on the run queue of the current Capability
void scheduleThread (Capability *cap, StgTSO *tso);
void awakenBlockedQueue (Capability *cap, StgTSO *tso);
#endif
+/* wakeUpRts()
+ *
+ * Causes an OS thread to wake up and run the scheduler, if necessary.
+ */
+void wakeUpRts(void);
+
/* unblockOne()
*
* Put the specified thread on the run queue of the given Capability.
* Called from STG : yes
* Locks assumed : we own the Capability.
*/
-StgTSO * unblockOne(Capability *cap, StgTSO *tso);
-
-/* raiseAsync()
- *
- * Raises an exception asynchronously in the specified thread.
- *
- * Called from STG : yes
- * Locks assumed : none
- */
-void raiseAsync(Capability *cap, StgTSO *tso, StgClosure *exception);
-
-/* suspendComputation()
- *
- * A variant of raiseAsync(), this strips the stack of the specified
- * thread down to the stop_here point, leaving a current closure on
- * top of the stack at [stop_here - 1].
- */
-void suspendComputation(Capability *cap, StgTSO *tso, StgPtr stop_here);
+StgTSO * unblockOne (Capability *cap, StgTSO *tso);
/* raiseExceptionHelper */
StgWord raiseExceptionHelper (StgRegTable *reg, StgTSO *tso, StgClosure *exception);
/* findRetryFrameHelper */
StgWord findRetryFrameHelper (StgTSO *tso);
-/* GetRoots(evac_fn f)
- *
- * Call f() for each root known to the scheduler.
- *
- * Called from STG : NO
- * Locks assumed : ????
- */
-void GetRoots(evac_fn);
-
/* workerStart()
*
* Entry point for a new worker task.
* Called from STG : NO
* Locks assumed : none
*/
-void workerStart(Task *task);
+#if defined(THREADED_RTS)
+void OSThreadProcAttr workerStart(Task *task);
+#endif
#if defined(GRAN)
void awaken_blocked_queue(StgBlockingQueueElement *q, StgClosure *node);
void initThread(StgTSO *tso, nat stack_size);
#endif
-/* Context switch flag.
- * Locks required : none (conflicts are harmless)
- */
-extern int RTS_VAR(context_switch);
-
/* The state of the scheduler. This is used to control the sequence
* of events during shutdown, and when the runtime is interrupted
* using ^C.
#define SCHED_INTERRUPTING 1 /* ^C detected, before threads are deleted */
#define SCHED_SHUTTING_DOWN 2 /* final shutdown */
-extern rtsBool RTS_VAR(sched_state);
+extern volatile StgWord RTS_VAR(sched_state);
/*
* flag that tracks whether we have done any execution in this time slice.
* INACTIVE to DONE_GC happens under sched_mutex. No lock required
* to set it to ACTIVITY_YES.
*/
-extern nat recent_activity;
+extern volatile StgWord recent_activity;
/* Thread queues.
* Locks required : sched_mutex
#endif
#endif
-/* Linked list of all threads.
- * Locks required : sched_mutex
- */
-extern StgTSO *RTS_VAR(all_threads);
-
/* Set to rtsTrue if there are threads on the blackhole_queue, and
* it is possible that one or more of them may be available to run.
* This flag is set to rtsFalse after we've checked the queue, and
*/
extern rtsBool blackholes_need_checking;
+extern rtsBool heap_overflow;
+
#if defined(THREADED_RTS)
extern Mutex RTS_VAR(sched_mutex);
#endif
-StgBool isThreadBound(StgTSO *tso);
-
SchedulerStatus rts_mainLazyIO(HaskellObj p, /*out*/HaskellObj *ret);
/* Called by shutdown_handler(). */
nat run_queue_len (void);
void resurrectThreads (StgTSO *);
+void performPendingThrowTos (StgTSO *);
void printAllThreads(void);
void print_bqe (StgBlockingQueueElement *bqe);
#endif
-void labelThread(StgPtr tso, char *label);
-
/* -----------------------------------------------------------------------------
* Some convenient macros/inline functions...
*/
* NOTE: tso->link should be END_TSO_QUEUE before calling this macro.
* ASSUMES: cap->running_task is the current task.
*/
-STATIC_INLINE void
+INLINE_HEADER void
appendToRunQueue (Capability *cap, StgTSO *tso)
{
- ASSERT(tso->link == END_TSO_QUEUE);
+ ASSERT(tso->_link == END_TSO_QUEUE);
if (cap->run_queue_hd == END_TSO_QUEUE) {
cap->run_queue_hd = tso;
} else {
- cap->run_queue_tl->link = tso;
+ setTSOLink(cap, cap->run_queue_tl, tso);
}
cap->run_queue_tl = tso;
+ postEvent (cap, EVENT_THREAD_RUNNABLE, tso->id, 0);
}
-/* Push a thread on the beginning of the run queue. Used for
- * newly awakened threads, so they get run as soon as possible.
+/* Push a thread on the beginning of the run queue.
* ASSUMES: cap->running_task is the current task.
*/
-STATIC_INLINE void
+INLINE_HEADER void
pushOnRunQueue (Capability *cap, StgTSO *tso)
{
- tso->link = cap->run_queue_hd;
+ setTSOLink(cap, tso, cap->run_queue_hd);
cap->run_queue_hd = tso;
if (cap->run_queue_tl == END_TSO_QUEUE) {
cap->run_queue_tl = tso;
/* Pop the first thread off the runnable queue.
*/
-STATIC_INLINE StgTSO *
+INLINE_HEADER StgTSO *
popRunQueue (Capability *cap)
{
StgTSO *t = cap->run_queue_hd;
ASSERT(t != END_TSO_QUEUE);
- cap->run_queue_hd = t->link;
- t->link = END_TSO_QUEUE;
+ cap->run_queue_hd = t->_link;
+ t->_link = END_TSO_QUEUE; // no write barrier req'd
if (cap->run_queue_hd == END_TSO_QUEUE) {
cap->run_queue_tl = END_TSO_QUEUE;
}
/* Add a thread to the end of the blocked queue.
*/
#if !defined(THREADED_RTS)
-STATIC_INLINE void
+INLINE_HEADER void
appendToBlockedQueue(StgTSO *tso)
{
- ASSERT(tso->link == END_TSO_QUEUE);
+ ASSERT(tso->_link == END_TSO_QUEUE);
if (blocked_queue_hd == END_TSO_QUEUE) {
blocked_queue_hd = tso;
} else {
- blocked_queue_tl->link = tso;
+ setTSOLink(&MainCapability, blocked_queue_tl, tso);
}
blocked_queue_tl = tso;
}
#endif
#if defined(THREADED_RTS)
-STATIC_INLINE void
-appendToWakeupQueue (Capability *cap, StgTSO *tso)
+// Assumes: my_cap is owned by the current Task. We hold
+// other_cap->lock, but we do not necessarily own other_cap; another
+// Task may be running on it.
+INLINE_HEADER void
+appendToWakeupQueue (Capability *my_cap, Capability *other_cap, StgTSO *tso)
{
- ASSERT(tso->link == END_TSO_QUEUE);
- if (cap->wakeup_queue_hd == END_TSO_QUEUE) {
- cap->wakeup_queue_hd = tso;
+ ASSERT(tso->_link == END_TSO_QUEUE);
+ if (other_cap->wakeup_queue_hd == END_TSO_QUEUE) {
+ other_cap->wakeup_queue_hd = tso;
} else {
- cap->wakeup_queue_tl->link = tso;
+ // my_cap is passed to setTSOLink() because it may need to
+ // write to the mutable list.
+ setTSOLink(my_cap, other_cap->wakeup_queue_tl, tso);
}
- cap->wakeup_queue_tl = tso;
+ other_cap->wakeup_queue_tl = tso;
}
#endif
/* Check whether various thread queues are empty
*/
-STATIC_INLINE rtsBool
+INLINE_HEADER rtsBool
emptyQueue (StgTSO *q)
{
return (q == END_TSO_QUEUE);
}
-STATIC_INLINE rtsBool
+INLINE_HEADER rtsBool
emptyRunQueue(Capability *cap)
{
return emptyQueue(cap->run_queue_hd);
}
#if defined(THREADED_RTS)
-STATIC_INLINE rtsBool
+INLINE_HEADER rtsBool
emptyWakeupQueue(Capability *cap)
{
return emptyQueue(cap->wakeup_queue_hd);
#define EMPTY_SLEEPING_QUEUE() (emptyQueue(sleeping_queue))
#endif
-STATIC_INLINE rtsBool
+INLINE_HEADER rtsBool
emptyThreadQueues(Capability *cap)
{
return emptyRunQueue(cap)
;
}
-#ifdef DEBUG
-void sched_belch(char *s, ...)
- GNU_ATTRIBUTE(format (printf, 1, 2));
-#endif
-
#endif /* !IN_STG_CODE */
-STATIC_INLINE void
-dirtyTSO (StgTSO *tso)
-{
- tso->flags |= TSO_DIRTY;
-}
-
#endif /* SCHEDULE_H */