! SPARC v8 __mpn_addmul_1 -- Multiply a limb vector with a limb and
! add the result to a second limb vector.
! Copyright (C) 1992-2016 Free Software Foundation, Inc.
! This file is part of the GNU MP Library.
! The GNU MP Library is free software; you can redistribute it and/or modify
! it under the terms of the GNU Lesser General Public License as published by
! the Free Software Foundation; either version 2.1 of the License, or (at your
! option) any later version.
! The GNU MP Library is distributed in the hope that it will be useful, but
! WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
! or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
! License for more details.
! You should have received a copy of the GNU Lesser General Public License
! along with the GNU MP Library; see the file COPYING.LIB. If not,
! see .
! INPUT PARAMETERS
! res_ptr o0
! s1_ptr o1
! size o2
! s2_limb o3
#include
ENTRY(__mpn_addmul_1)
ld [%o1+0],%o4 ! 1
sll %o2,4,%g1
orcc %g0,%g0,%g2
mov %o7,%g4 ! Save return address register
and %g1,(4-1)<<4,%g1
1: call 2f
add %o7,3f-1b,%g3
2: jmp %g3+%g1
mov %g4,%o7 ! Restore return address register
.align 4
3:
LOC(00):
add %o0,-4,%o0
b LOC(loop00) /* 4, 8, 12, ... */
add %o1,-4,%o1
nop
LOC(01):
b LOC(loop01) /* 1, 5, 9, ... */
nop
nop
nop
LOC(10):
add %o0,-12,%o0 /* 2, 6, 10, ... */
b LOC(loop10)
add %o1,4,%o1
nop
LOC(11):
add %o0,-8,%o0 /* 3, 7, 11, ... */
b LOC(loop11)
add %o1,-8,%o1
nop
LOC(loop):
addcc %g3,%g2,%g3 ! 1
ld [%o1+4],%o4 ! 2
rd %y,%g2 ! 1
addx %g0,%g2,%g2
ld [%o0+0],%g1 ! 2
addcc %g1,%g3,%g3
st %g3,[%o0+0] ! 1
LOC(loop00):
umul %o4,%o3,%g3 ! 2
ld [%o0+4],%g1 ! 2
addxcc %g3,%g2,%g3 ! 2
ld [%o1+8],%o4 ! 3
rd %y,%g2 ! 2
addx %g0,%g2,%g2
nop
addcc %g1,%g3,%g3
st %g3,[%o0+4] ! 2
LOC(loop11):
umul %o4,%o3,%g3 ! 3
addxcc %g3,%g2,%g3 ! 3
ld [%o1+12],%o4 ! 4
rd %y,%g2 ! 3
add %o1,16,%o1
addx %g0,%g2,%g2
ld [%o0+8],%g1 ! 2
addcc %g1,%g3,%g3
st %g3,[%o0+8] ! 3
LOC(loop10):
umul %o4,%o3,%g3 ! 4
addxcc %g3,%g2,%g3 ! 4
ld [%o1+0],%o4 ! 1
rd %y,%g2 ! 4
addx %g0,%g2,%g2
ld [%o0+12],%g1 ! 2
addcc %g1,%g3,%g3
st %g3,[%o0+12] ! 4
add %o0,16,%o0
addx %g0,%g2,%g2
LOC(loop01):
addcc %o2,-4,%o2
bg LOC(loop)
umul %o4,%o3,%g3 ! 1
addcc %g3,%g2,%g3 ! 4
rd %y,%g2 ! 4
addx %g0,%g2,%g2
ld [%o0+0],%g1 ! 2
addcc %g1,%g3,%g3
st %g3,[%o0+0] ! 4
retl
addx %g0,%g2,%o0
END(__mpn_addmul_1)