projects
/
ghc-hetmet.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Remove vectored returns.
[ghc-hetmet.git]
/
rts
/
sm
/
Compact.c
diff --git
a/rts/sm/Compact.c
b/rts/sm/Compact.c
index
844b770
..
feebef8
100644
(file)
--- a/
rts/sm/Compact.c
+++ b/
rts/sm/Compact.c
@@
-16,8
+16,6
@@
#include "RtsUtils.h"
#include "RtsFlags.h"
#include "OSThreads.h"
#include "RtsUtils.h"
#include "RtsFlags.h"
#include "OSThreads.h"
-#include "Storage.h"
-#include "Stable.h"
#include "BlockAlloc.h"
#include "MBlock.h"
#include "GC.h"
#include "BlockAlloc.h"
#include "MBlock.h"
#include "GC.h"
@@
-268,7
+266,6
@@
thread_stack(StgPtr p, StgPtr stack_end)
case STOP_FRAME:
case CATCH_FRAME:
case RET_SMALL:
case STOP_FRAME:
case CATCH_FRAME:
case RET_SMALL:
- case RET_VEC_SMALL:
bitmap = BITMAP_BITS(info->i.layout.bitmap);
size = BITMAP_SIZE(info->i.layout.bitmap);
p++;
bitmap = BITMAP_BITS(info->i.layout.bitmap);
size = BITMAP_SIZE(info->i.layout.bitmap);
p++;
@@
-300,7
+297,6
@@
thread_stack(StgPtr p, StgPtr stack_end)
// large bitmap (> 32 entries, or 64 on a 64-bit machine)
case RET_BIG:
// large bitmap (> 32 entries, or 64 on a 64-bit machine)
case RET_BIG:
- case RET_VEC_BIG:
p++;
size = GET_LARGE_BITMAP(&info->i)->size;
thread_large_bitmap(p, GET_LARGE_BITMAP(&info->i), size);
p++;
size = GET_LARGE_BITMAP(&info->i)->size;
thread_large_bitmap(p, GET_LARGE_BITMAP(&info->i), size);
@@
-532,7
+528,6
@@
thread_obj (StgInfoTable *info, StgPtr p)
thread_(&bco->instrs);
thread_(&bco->literals);
thread_(&bco->ptrs);
thread_(&bco->instrs);
thread_(&bco->literals);
thread_(&bco->ptrs);
- thread_(&bco->itbls);
return p + bco_sizeW(bco);
}
return p + bco_sizeW(bco);
}
@@
-768,6
+763,10
@@
update_fwd_compact( bdescr *blocks )
// know the destination without the size, because we may
// spill into the next block. So we have to run down the
// threaded list and get the info ptr first.
// know the destination without the size, because we may
// spill into the next block. So we have to run down the
// threaded list and get the info ptr first.
+ //
+ // ToDo: one possible avenue of attack is to use the fact
+ // that if (p&BLOCK_MASK) >= (free&BLOCK_MASK), then we
+ // definitely have enough room. Also see bug #1147.
info = get_threaded_info(p);
q = p;
info = get_threaded_info(p);
q = p;