X-Git-Url: http://git.megacz.com/?a=blobdiff_plain;f=rts%2Fgmp%2Fmpn%2Fsparc32%2Fv9%2Fmul_1.asm;fp=rts%2Fgmp%2Fmpn%2Fsparc32%2Fv9%2Fmul_1.asm;h=0000000000000000000000000000000000000000;hb=fdf1cd0399158308769fcb2ab7e46e215a68d865;hp=f8f0fdd8c2727bc0cec997b26e20efe486480d96;hpb=e552cfc427d2734b9a9629f2ab1d22f493e775f6;p=ghc-hetmet.git diff --git a/rts/gmp/mpn/sparc32/v9/mul_1.asm b/rts/gmp/mpn/sparc32/v9/mul_1.asm deleted file mode 100644 index f8f0fdd..0000000 --- a/rts/gmp/mpn/sparc32/v9/mul_1.asm +++ /dev/null @@ -1,267 +0,0 @@ -dnl SPARC v9 32-bit mpn_mul_1 -- Multiply a limb vector with a limb and -dnl store the result in a second limb vector. - -dnl Copyright (C) 1998, 2000 Free Software Foundation, Inc. - -dnl This file is part of the GNU MP Library. - -dnl The GNU MP Library is free software; you can redistribute it and/or modify -dnl it under the terms of the GNU Lesser General Public License as published -dnl by the Free Software Foundation; either version 2.1 of the License, or (at -dnl your option) any later version. - -dnl The GNU MP Library is distributed in the hope that it will be useful, but -dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY -dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public -dnl License for more details. - -dnl You should have received a copy of the GNU Lesser General Public License -dnl along with the GNU MP Library; see the file COPYING.LIB. If not, write to -dnl the Free Software Foundation, Inc., 59 Temple Place - Suite 330, Boston, -dnl MA 02111-1307, USA. - - -include(`../config.m4') - -C INPUT PARAMETERS -C res_ptr i0 -C s1_ptr i1 -C size i2 -C s2_limb i3 - -ASM_START() - - TEXT - ALIGN(4) -L(noll): - .word 0 - -PROLOGUE(mpn_mul_1) - save %sp,-256,%sp - -ifdef(`PIC', -`L(pc): rd %pc,%o7 - ld [%o7+L(noll)-L(pc)],%f10', -` sethi %hi(L(noll)),%g1 - ld [%g1+%lo(L(noll))],%f10') - - sethi %hi(0xffff0000),%o0 - andn %i3,%o0,%o0 - st %o0,[%fp-16] - ld [%fp-16],%f11 - fxtod %f10,%f6 - - srl %i3,16,%o0 - st %o0,[%fp-16] - ld [%fp-16],%f11 - fxtod %f10,%f8 - - mov 0,%g3 C cy = 0 - - ld [%i1],%f11 - subcc %i2,1,%i2 - be,pn %icc,L(end1) - add %i1,4,%i1 C s1_ptr++ - - fxtod %f10,%f2 - ld [%i1],%f11 - add %i1,4,%i1 C s1_ptr++ - fmuld %f2,%f8,%f16 - fmuld %f2,%f6,%f4 - fdtox %f16,%f14 - std %f14,[%fp-24] - fdtox %f4,%f12 - subcc %i2,1,%i2 - be,pn %icc,L(end2) - std %f12,[%fp-16] - - fxtod %f10,%f2 - ld [%i1],%f11 - add %i1,4,%i1 C s1_ptr++ - fmuld %f2,%f8,%f16 - fmuld %f2,%f6,%f4 - fdtox %f16,%f14 - std %f14,[%fp-40] - fdtox %f4,%f12 - subcc %i2,1,%i2 - be,pn %icc,L(end3) - std %f12,[%fp-32] - - fxtod %f10,%f2 - ld [%i1],%f11 - add %i1,4,%i1 C s1_ptr++ - ldx [%fp-24],%g2 C p16 - fmuld %f2,%f8,%f16 - ldx [%fp-16],%g1 C p0 - fmuld %f2,%f6,%f4 - sllx %g2,16,%g2 C align p16 - fdtox %f16,%f14 - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - std %f14,[%fp-24] - fdtox %f4,%f12 - add %i0,4,%i0 C res_ptr++ - subcc %i2,1,%i2 - be,pn %icc,L(end4) - std %f12,[%fp-16] - - b,a L(loopm) - - .align 16 -C BEGIN LOOP -L(loop): - fxtod %f10,%f2 - ld [%i1],%f11 - add %i1,4,%i1 C s1_ptr++ - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-24],%g2 C p16 - fmuld %f2,%f8,%f16 - ldx [%fp-16],%g1 C p0 - fmuld %f2,%f6,%f4 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - fdtox %f16,%f14 - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - std %f14,[%fp-24] - fdtox %f4,%f12 - std %f12,[%fp-16] - subcc %i2,1,%i2 - be,pn %icc,L(loope) - add %i0,4,%i0 C res_ptr++ -L(loopm): - fxtod %f10,%f2 - ld [%i1],%f11 - add %i1,4,%i1 C s1_ptr++ - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-40],%g2 C p16 - fmuld %f2,%f8,%f16 - ldx [%fp-32],%g1 C p0 - fmuld %f2,%f6,%f4 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - fdtox %f16,%f14 - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - std %f14,[%fp-40] - fdtox %f4,%f12 - std %f12,[%fp-32] - subcc %i2,1,%i2 - bne,pt %icc,L(loop) - add %i0,4,%i0 C res_ptr++ -C END LOOP - - fxtod %f10,%f2 - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-24],%g2 C p16 - fmuld %f2,%f8,%f16 - ldx [%fp-16],%g1 C p0 - fmuld %f2,%f6,%f4 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - b,a L(xxx) -L(loope): -L(end4): - fxtod %f10,%f2 - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-40],%g2 C p16 - fmuld %f2,%f8,%f16 - ldx [%fp-32],%g1 C p0 - fmuld %f2,%f6,%f4 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - fdtox %f16,%f14 - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - std %f14,[%fp-40] - fdtox %f4,%f12 - std %f12,[%fp-32] - add %i0,4,%i0 C res_ptr++ - - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-24],%g2 C p16 - ldx [%fp-16],%g1 C p0 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - b,a L(yyy) - -L(end3): - fxtod %f10,%f2 - ldx [%fp-24],%g2 C p16 - fmuld %f2,%f8,%f16 - ldx [%fp-16],%g1 C p0 - fmuld %f2,%f6,%f4 - sllx %g2,16,%g2 C align p16 -L(xxx): fdtox %f16,%f14 - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - std %f14,[%fp-24] - fdtox %f4,%f12 - std %f12,[%fp-16] - add %i0,4,%i0 C res_ptr++ - - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-40],%g2 C p16 - ldx [%fp-32],%g1 C p0 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - add %i0,4,%i0 C res_ptr++ - - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-24],%g2 C p16 - ldx [%fp-16],%g1 C p0 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - add %i0,4,%i0 C res_ptr++ - b,a L(ret) - -L(end2): - fxtod %f10,%f2 - fmuld %f2,%f8,%f16 - fmuld %f2,%f6,%f4 - fdtox %f16,%f14 - std %f14,[%fp-40] - fdtox %f4,%f12 - std %f12,[%fp-32] - ldx [%fp-24],%g2 C p16 - ldx [%fp-16],%g1 C p0 - sllx %g2,16,%g2 C align p16 -L(yyy): add %g2,%g1,%g1 C add p16 to p0 (ADD1) - add %i0,4,%i0 C res_ptr++ - - add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - ldx [%fp-40],%g2 C p16 - ldx [%fp-32],%g1 C p0 - sllx %g2,16,%g2 C align p16 - st %g4,[%i0-4] - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - add %i0,4,%i0 C res_ptr++ - b,a L(ret) - -L(end1): - fxtod %f10,%f2 - fmuld %f2,%f8,%f16 - fmuld %f2,%f6,%f4 - fdtox %f16,%f14 - std %f14,[%fp-24] - fdtox %f4,%f12 - std %f12,[%fp-16] - - ldx [%fp-24],%g2 C p16 - ldx [%fp-16],%g1 C p0 - sllx %g2,16,%g2 C align p16 - add %g2,%g1,%g1 C add p16 to p0 (ADD1) - add %i0,4,%i0 C res_ptr++ - -L(ret): add %g3,%g1,%g4 C p += cy - srlx %g4,32,%g3 - st %g4,[%i0-4] - - ret - restore %g0,%g3,%o0 C sideeffect: put cy in retreg -EPILOGUE(mpn_mul_1)