dnl SPARC v9 32-bit mpn_addmul_1 -- Multiply a limb vector with a limb and dnl add the result to a second limb vector. dnl Copyright (C) 1998, 2000 Free Software Foundation, Inc. dnl This file is part of the GNU MP Library. dnl The GNU MP Library is free software; you can redistribute it and/or modify dnl it under the terms of the GNU Library General Public License as published dnl by the Free Software Foundation; either version 2 of the License, or (at dnl your option) any later version. dnl The GNU MP Library is distributed in the hope that it will be useful, but dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Library General Public dnl License for more details. dnl You should have received a copy of the GNU Library General Public License dnl along with the GNU MP Library; see the file COPYING.LIB. If not, write to dnl the Free Software Foundation, Inc., 59 Temple Place - Suite 330, Boston, dnl MA 02111-1307, USA. include(`../config.m4') C INPUT PARAMETERS C res_ptr i0 C s1_ptr i1 C size i2 C s2_limb i3 ASM_START() TEXT ALIGN(4) L(noll): .word 0 PROLOGUE(mpn_addmul_1) save %sp,-256,%sp ifdef(`PIC', `L(pc): rd %pc,%o7 ld [%o7+L(noll)-L(pc)],%f10', ` sethi %hi(L(noll)),%g1 ld [%g1+%lo(L(noll))],%f10') sethi %hi(0xffff0000),%o0 andn %i3,%o0,%o0 st %o0,[%fp-16] ld [%fp-16],%f11 fxtod %f10,%f6 srl %i3,16,%o0 st %o0,[%fp-16] ld [%fp-16],%f11 fxtod %f10,%f8 mov 0,%g3 C cy = 0 ld [%i1],%f11 subcc %i2,1,%i2 be,pn %icc,L(end1) add %i1,4,%i1 C s1_ptr++ fxtod %f10,%f2 ld [%i1],%f11 add %i1,4,%i1 C s1_ptr++ fmuld %f2,%f8,%f16 fmuld %f2,%f6,%f4 fdtox %f16,%f14 std %f14,[%fp-24] fdtox %f4,%f12 subcc %i2,1,%i2 be,pn %icc,L(end2) std %f12,[%fp-16] fxtod %f10,%f2 ld [%i1],%f11 add %i1,4,%i1 C s1_ptr++ fmuld %f2,%f8,%f16 fmuld %f2,%f6,%f4 fdtox %f16,%f14 std %f14,[%fp-40] fdtox %f4,%f12 subcc %i2,1,%i2 be,pn %icc,L(end3) std %f12,[%fp-32] fxtod %f10,%f2 ld [%i1],%f11 add %i1,4,%i1 C s1_ptr++ ld [%i0],%g5 ldx [%fp-24],%g2 C p16 fmuld %f2,%f8,%f16 ldx [%fp-16],%g1 C p0 fmuld %f2,%f6,%f4 sllx %g2,16,%g2 C align p16 fdtox %f16,%f14 add %g2,%g1,%g1 C add p16 to p0 (ADD1) std %f14,[%fp-24] fdtox %f4,%f12 add %i0,4,%i0 C res_ptr++ subcc %i2,1,%i2 be,pn %icc,L(end4) std %f12,[%fp-16] b,a L(loopm) .align 16 C BEGIN LOOP L(loop): fxtod %f10,%f2 ld [%i1],%f11 add %i1,4,%i1 C s1_ptr++ add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-24],%g2 C p16 fmuld %f2,%f8,%f16 ldx [%fp-16],%g1 C p0 fmuld %f2,%f6,%f4 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] fdtox %f16,%f14 add %g2,%g1,%g1 C add p16 to p0 (ADD1) std %f14,[%fp-24] fdtox %f4,%f12 std %f12,[%fp-16] subcc %i2,1,%i2 be,pn %icc,L(loope) add %i0,4,%i0 C res_ptr++ L(loopm): fxtod %f10,%f2 ld [%i1],%f11 add %i1,4,%i1 C s1_ptr++ add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-40],%g2 C p16 fmuld %f2,%f8,%f16 ldx [%fp-32],%g1 C p0 fmuld %f2,%f6,%f4 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] fdtox %f16,%f14 add %g2,%g1,%g1 C add p16 to p0 (ADD1) std %f14,[%fp-40] fdtox %f4,%f12 std %f12,[%fp-32] subcc %i2,1,%i2 bne,pt %icc,L(loop) add %i0,4,%i0 C res_ptr++ C END LOOP fxtod %f10,%f2 add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-24],%g2 C p16 fmuld %f2,%f8,%f16 ldx [%fp-16],%g1 C p0 fmuld %f2,%f6,%f4 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] b,a L(xxx) L(loope): L(end4): fxtod %f10,%f2 add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-40],%g2 C p16 fmuld %f2,%f8,%f16 ldx [%fp-32],%g1 C p0 fmuld %f2,%f6,%f4 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] fdtox %f16,%f14 add %g2,%g1,%g1 C add p16 to p0 (ADD1) std %f14,[%fp-40] fdtox %f4,%f12 std %f12,[%fp-32] add %i0,4,%i0 C res_ptr++ add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-24],%g2 C p16 ldx [%fp-16],%g1 C p0 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] b,a L(yyy) L(end3): fxtod %f10,%f2 ld [%i0],%g5 ldx [%fp-24],%g2 C p16 fmuld %f2,%f8,%f16 ldx [%fp-16],%g1 C p0 fmuld %f2,%f6,%f4 sllx %g2,16,%g2 C align p16 L(xxx): fdtox %f16,%f14 add %g2,%g1,%g1 C add p16 to p0 (ADD1) std %f14,[%fp-24] fdtox %f4,%f12 std %f12,[%fp-16] add %i0,4,%i0 C res_ptr++ add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-40],%g2 C p16 ldx [%fp-32],%g1 C p0 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] add %g2,%g1,%g1 C add p16 to p0 (ADD1) add %i0,4,%i0 C res_ptr++ add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-24],%g2 C p16 ldx [%fp-16],%g1 C p0 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] add %g2,%g1,%g1 C add p16 to p0 (ADD1) add %i0,4,%i0 C res_ptr++ b,a L(ret) L(end2): fxtod %f10,%f2 fmuld %f2,%f8,%f16 fmuld %f2,%f6,%f4 fdtox %f16,%f14 std %f14,[%fp-40] fdtox %f4,%f12 std %f12,[%fp-32] ld [%i0],%g5 ldx [%fp-24],%g2 C p16 ldx [%fp-16],%g1 C p0 sllx %g2,16,%g2 C align p16 L(yyy): add %g2,%g1,%g1 C add p16 to p0 (ADD1) add %i0,4,%i0 C res_ptr++ add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy ld [%i0],%g5 srlx %g4,32,%g3 ldx [%fp-40],%g2 C p16 ldx [%fp-32],%g1 C p0 sllx %g2,16,%g2 C align p16 st %g4,[%i0-4] add %g2,%g1,%g1 C add p16 to p0 (ADD1) add %i0,4,%i0 C res_ptr++ b,a L(ret) L(end1): fxtod %f10,%f2 fmuld %f2,%f8,%f16 fmuld %f2,%f6,%f4 fdtox %f16,%f14 std %f14,[%fp-24] fdtox %f4,%f12 std %f12,[%fp-16] ld [%i0],%g5 ldx [%fp-24],%g2 C p16 ldx [%fp-16],%g1 C p0 sllx %g2,16,%g2 C align p16 add %g2,%g1,%g1 C add p16 to p0 (ADD1) add %i0,4,%i0 C res_ptr++ L(ret): add %g5,%g1,%g1 C add *res_ptr to p0 (ADD2) add %g3,%g1,%g4 C p += cy srlx %g4,32,%g3 st %g4,[%i0-4] ret restore %g0,%g3,%o0 C sideeffect: put cy in retreg EPILOGUE(mpn_addmul_1)