import ZipCfgCmmRep (Convention(..))
import Constants
+import qualified Data.List as L
import StaticFlags (opt_Unregisterised)
import Outputable
-import Panic
-- Calculate the 'GlobalReg' or stack locations for function call
-- parameters as used by the Cmm calling convention.
-- Stack parameters are returned as word offsets.
assignArguments :: (a -> CmmType) -> [a] -> ArgumentFormat a WordOff
-assignArguments f reps = assignments
+assignArguments f reps = panic "assignArguments only used in dead codegen" -- assignments
where
availRegs = getRegsWithNode
(sizes, assignments) = unzip $ assignArguments' reps (negate (sum sizes)) availRegs
-- | JD: For the new stack story, I want arguments passed on the stack to manifest as
-- positive offsets in a CallArea, not negative offsets from the stack pointer.
-- Also, I want byte offsets, not word offsets.
--- The first argument tells us whether we are assigning positions for call arguments
--- or return results. The distinction matters because some conventions use different
--- global registers in each case. In particular, the native calling convention
--- uses the `node' register to pass the closure environment.
-assignArgumentsPos :: (Outputable a) => Convention -> Bool -> (a -> CmmType) -> [a] ->
+assignArgumentsPos :: (Outputable a) => Convention -> (a -> CmmType) -> [a] ->
ArgumentFormat a ByteOff
-assignArgumentsPos conv isCall arg_ty reps = map cvt assignments
+assignArgumentsPos conv arg_ty reps = assignments -- old_assts'
where -- The calling conventions (CgCallConv.hs) are complicated, to say the least
- regs = if isCall then
- case (reps, conv) of
- (_, Native) -> getRegsWithoutNode
- (_, GC ) -> getRegsWithNode
- (_, PrimOp) -> allRegs
- (_, Slow ) -> noRegs
- (_, _ ) -> getRegsWithoutNode
- else
- case (reps, conv) of
- ([_], _) -> allRegs
- (_, Native) -> getRegsWithNode
- (_, GC ) -> getRegsWithNode
- (_, PrimOp) -> getRegsWithNode
- (_, Slow ) -> noRegs
- (_, _ ) -> getRegsWithNode
- (sizes, assignments) = unzip $ assignArguments' reps (sum sizes) regs
+ regs = case (reps, conv) of
+ (_, NativeNodeCall) -> getRegsWithNode
+ (_, NativeDirectCall) -> getRegsWithoutNode
+ ([_], NativeReturn) -> allRegs
+ (_, NativeReturn) -> getRegsWithNode
+ (_, GC) -> getRegsWithNode
+ (_, PrimOpCall) -> allRegs
+ ([_], PrimOpReturn) -> allRegs
+ (_, PrimOpReturn) -> getRegsWithNode
+ (_, Slow) -> noRegs
+ _ -> pprPanic "Unknown calling convention" (ppr conv)
+ -- The calling conventions first assign arguments to registers,
+ -- then switch to the stack when we first run out of registers
+ -- (even if there are still available registers for args of a
+ -- different type).
+ -- When returning an unboxed tuple, we also separate the stack
+ -- arguments by pointerhood.
+ (reg_assts, stk_args) = assign_regs [] reps regs
+ stk_args' = case conv of NativeReturn -> part
+ PrimOpReturn -> part
+ _ -> stk_args
+ where part = uncurry (++)
+ (L.partition (not . isGcPtrType . arg_ty) stk_args)
+ stk_assts = assign_stk 0 [] (reverse stk_args')
+ assignments = reg_assts ++ stk_assts
+
+ assign_regs assts [] _ = (assts, [])
+ assign_regs assts (r:rs) regs = if isFloatType ty then float else int
+ where float = case (w, regs) of
+ (W32, (vs, f:fs, ds, ls)) -> k (RegisterParam f, (vs, fs, ds, ls))
+ (W64, (vs, fs, d:ds, ls)) -> k (RegisterParam d, (vs, fs, ds, ls))
+ (W80, _) -> panic "F80 unsupported register type"
+ _ -> (assts, (r:rs))
+ int = case (w, regs) of
+ (W128, _) -> panic "W128 unsupported register type"
+ (_, (v:vs, fs, ds, ls)) | widthInBits w <= widthInBits wordWidth
+ -> k (RegisterParam (v gcp), (vs, fs, ds, ls))
+ (_, (vs, fs, ds, l:ls)) | widthInBits w > widthInBits wordWidth
+ -> k (RegisterParam l, (vs, fs, ds, ls))
+ _ -> (assts, (r:rs))
+ k (asst, regs') = assign_regs ((r, asst) : assts) rs regs'
+ ty = arg_ty r
+ w = typeWidth ty
+ gcp | isGcPtrType ty = VGcPtr
+ | otherwise = VNonGcPtr
+
+ assign_stk offset assts [] = assts
+ assign_stk offset assts (r:rs) = assign_stk off' ((r, StackParam off') : assts) rs
+ where w = typeWidth (arg_ty r)
+ size = (((widthInBytes w - 1) `div` wORD_SIZE) + 1) * wORD_SIZE
+ off' = offset + size
+
+
+ -- DEAD CODE:
+ (old_sizes, old_assignments) = unzip $ assignArguments' reps (sum old_sizes) regs
+ old_assts' = map cvt old_assignments
+
assignArguments' [] _ _ = []
assignArguments' (r:rs) offset avails =
(size, (r,assignment)):assignArguments' rs new_offset remaining
-- Assigning a slot using negative offsets from the stack pointer.
-- JD: I don't know why this convention stops using all the registers
--- after running out of one class of registers.
+-- after running out of one class of registers, but that's how it is.
assign_slot_neg :: SlotAssigner
assign_slot_neg width off _regs =
(StackParam $ off, off + size, size, ([], [], [], [])) where size = slot_size' width
assign_bits_reg _ W128 _ _ _ = panic "W128 is not a supported register type"
assign_bits_reg _ w off gcp (v:vs, fs, ds, ls)
| widthInBits w <= widthInBits wordWidth =
- pprTrace "long regs" (ppr ls <+> ppr wordWidth <+> ppr mAX_Real_Long_REG) $ (RegisterParam (v gcp), off, 0, (vs, fs, ds, ls))
+ (RegisterParam (v gcp), off, 0, (vs, fs, ds, ls))
assign_bits_reg _ w off _ (vs, fs, ds, l:ls)
| widthInBits w > widthInBits wordWidth =
- pprTrace "long regs" (ppr ls <+> ppr wordWidth <+> ppr mAX_Real_Long_REG) $ (RegisterParam l, off, 0, (vs, fs, ds, ls))
-assign_bits_reg assign_slot w off _ regs@(_, _, _, ls) =
- pprTrace "long regs" (ppr w <+> ppr ls <+> ppr wordWidth <+> ppr mAX_Real_Long_REG <+> ppr mAX_Long_REG) $ assign_slot w off regs
+ (RegisterParam l, off, 0, (vs, fs, ds, ls))
+assign_bits_reg assign_slot w off _ regs@(_, _, _, _) = assign_slot w off regs
assign_float_reg :: SlotAssigner -> Width -> WordOff -> AvailRegs -> Assignment
assign_float_reg _ W32 off (vs, f:fs, ds, ls) = (RegisterParam $ f, off, 0, (vs, fs, ds, ls))