diff options
Diffstat (limited to 'sysdeps/sparc/sparc32/sub_n.S')
-rw-r--r-- | sysdeps/sparc/sparc32/sub_n.S | 406 |
1 files changed, 212 insertions, 194 deletions
diff --git a/sysdeps/sparc/sparc32/sub_n.S b/sysdeps/sparc/sparc32/sub_n.S index b7a11958e2..74400600bd 100644 --- a/sysdeps/sparc/sparc32/sub_n.S +++ b/sysdeps/sparc/sparc32/sub_n.S @@ -1,20 +1,20 @@ ! SPARC __mpn_sub_n -- Subtract two limb vectors of the same length > 0 and ! store difference in a third limb vector. - +! ! Copyright (C) 1995, 1996 Free Software Foundation, Inc. - +! ! This file is part of the GNU MP Library. - +! ! The GNU MP Library is free software; you can redistribute it and/or modify ! it under the terms of the GNU Library General Public License as published by ! the Free Software Foundation; either version 2 of the License, or (at your ! option) any later version. - +! ! The GNU MP Library is distributed in the hope that it will be useful, but ! WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY ! or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Library General Public ! License for more details. - +! ! You should have received a copy of the GNU Library General Public License ! along with the GNU MP Library; see the file COPYING.LIB. If not, write to ! the Free Software Foundation, Inc., 59 Temple Place - Suite 330, Boston, @@ -22,290 +22,308 @@ ! INPUT PARAMETERS -#define res_ptr %o0 -#define s1_ptr %o1 -#define s2_ptr %o2 -#define size %o3 +#define RES_PTR %o0 +#define S1_PTR %o1 +#define S2_PTR %o2 +#define SIZE %o3 -#include "sysdep.h" +#include <sysdep.h> - .text - .align 4 - .global C_SYMBOL_NAME(__mpn_sub_n) -C_SYMBOL_NAME(__mpn_sub_n): - xor s2_ptr,res_ptr,%g1 +ENTRY(__mpn_sub_n) + xor S2_PTR,RES_PTR,%g1 andcc %g1,4,%g0 - bne L1 ! branch if alignment differs + bne LOC(1) ! branch if alignment differs nop ! ** V1a ** - andcc res_ptr,4,%g0 ! res_ptr unaligned? Side effect: cy=0 - be L_v1 ! if no, branch + andcc RES_PTR,4,%g0 ! RES_PTR unaligned? Side effect: cy=0 + be LOC(v1) ! if no, branch nop -/* Add least significant limb separately to align res_ptr and s2_ptr */ - ld [s1_ptr],%g4 - add s1_ptr,4,s1_ptr - ld [s2_ptr],%g2 - add s2_ptr,4,s2_ptr - add size,-1,size +/* Add least significant limb separately to align RES_PTR and S2_PTR */ + ld [S1_PTR],%g4 + add S1_PTR,4,S1_PTR + ld [S2_PTR],%g2 + add S2_PTR,4,S2_PTR + add SIZE,-1,SIZE subcc %g4,%g2,%o4 - st %o4,[res_ptr] - add res_ptr,4,res_ptr -L_v1: addx %g0,%g0,%o4 ! save cy in register - cmp size,2 ! if size < 2 ... - bl Lend2 ! ... branch to tail code + st %o4,[RES_PTR] + add RES_PTR,4,RES_PTR +LOC(v1): + addx %g0,%g0,%o4 ! save cy in register + cmp SIZE,2 ! if SIZE < 2 ... + bl LOC(end2) ! ... branch to tail code subcc %g0,%o4,%g0 ! restore cy - ld [s1_ptr+0],%g4 - addcc size,-10,size - ld [s1_ptr+4],%g1 - ldd [s2_ptr+0],%g2 - blt Lfin1 + ld [S1_PTR+0],%g4 + addcc SIZE,-10,SIZE + ld [S1_PTR+4],%g1 + ldd [S2_PTR+0],%g2 + blt LOC(fin1) subcc %g0,%o4,%g0 ! restore cy /* Add blocks of 8 limbs until less than 8 limbs remain */ -Loop1: subxcc %g4,%g2,%o4 - ld [s1_ptr+8],%g4 +LOC(loop1): + subxcc %g4,%g2,%o4 + ld [S1_PTR+8],%g4 subxcc %g1,%g3,%o5 - ld [s1_ptr+12],%g1 - ldd [s2_ptr+8],%g2 - std %o4,[res_ptr+0] + ld [S1_PTR+12],%g1 + ldd [S2_PTR+8],%g2 + std %o4,[RES_PTR+0] subxcc %g4,%g2,%o4 - ld [s1_ptr+16],%g4 + ld [S1_PTR+16],%g4 subxcc %g1,%g3,%o5 - ld [s1_ptr+20],%g1 - ldd [s2_ptr+16],%g2 - std %o4,[res_ptr+8] + ld [S1_PTR+20],%g1 + ldd [S2_PTR+16],%g2 + std %o4,[RES_PTR+8] subxcc %g4,%g2,%o4 - ld [s1_ptr+24],%g4 + ld [S1_PTR+24],%g4 subxcc %g1,%g3,%o5 - ld [s1_ptr+28],%g1 - ldd [s2_ptr+24],%g2 - std %o4,[res_ptr+16] + ld [S1_PTR+28],%g1 + ldd [S2_PTR+24],%g2 + std %o4,[RES_PTR+16] subxcc %g4,%g2,%o4 - ld [s1_ptr+32],%g4 + ld [S1_PTR+32],%g4 subxcc %g1,%g3,%o5 - ld [s1_ptr+36],%g1 - ldd [s2_ptr+32],%g2 - std %o4,[res_ptr+24] + ld [S1_PTR+36],%g1 + ldd [S2_PTR+32],%g2 + std %o4,[RES_PTR+24] addx %g0,%g0,%o4 ! save cy in register - addcc size,-8,size - add s1_ptr,32,s1_ptr - add s2_ptr,32,s2_ptr - add res_ptr,32,res_ptr - bge Loop1 + addcc SIZE,-8,SIZE + add S1_PTR,32,S1_PTR + add S2_PTR,32,S2_PTR + add RES_PTR,32,RES_PTR + bge LOC(loop1) subcc %g0,%o4,%g0 ! restore cy -Lfin1: addcc size,8-2,size - blt Lend1 +LOC(fin1): + addcc SIZE,8-2,SIZE + blt LOC(end1) subcc %g0,%o4,%g0 ! restore cy /* Add blocks of 2 limbs until less than 2 limbs remain */ -Loope1: subxcc %g4,%g2,%o4 - ld [s1_ptr+8],%g4 +LOC(loope1): + subxcc %g4,%g2,%o4 + ld [S1_PTR+8],%g4 subxcc %g1,%g3,%o5 - ld [s1_ptr+12],%g1 - ldd [s2_ptr+8],%g2 - std %o4,[res_ptr+0] + ld [S1_PTR+12],%g1 + ldd [S2_PTR+8],%g2 + std %o4,[RES_PTR+0] addx %g0,%g0,%o4 ! save cy in register - addcc size,-2,size - add s1_ptr,8,s1_ptr - add s2_ptr,8,s2_ptr - add res_ptr,8,res_ptr - bge Loope1 + addcc SIZE,-2,SIZE + add S1_PTR,8,S1_PTR + add S2_PTR,8,S2_PTR + add RES_PTR,8,RES_PTR + bge LOC(loope1) subcc %g0,%o4,%g0 ! restore cy -Lend1: subxcc %g4,%g2,%o4 +LOC(end1): + subxcc %g4,%g2,%o4 subxcc %g1,%g3,%o5 - std %o4,[res_ptr+0] + std %o4,[RES_PTR+0] addx %g0,%g0,%o4 ! save cy in register - andcc size,1,%g0 - be Lret1 + andcc SIZE,1,%g0 + be LOC(ret1) subcc %g0,%o4,%g0 ! restore cy /* Add last limb */ - ld [s1_ptr+8],%g4 - ld [s2_ptr+8],%g2 + ld [S1_PTR+8],%g4 + ld [S2_PTR+8],%g2 subxcc %g4,%g2,%o4 - st %o4,[res_ptr+8] + st %o4,[RES_PTR+8] -Lret1: retl +LOC(ret1): + retl addx %g0,%g0,%o0 ! return carry-out from most sign. limb -L1: xor s1_ptr,res_ptr,%g1 +LOC(1): xor S1_PTR,RES_PTR,%g1 andcc %g1,4,%g0 - bne L2 + bne LOC(2) nop ! ** V1b ** - andcc res_ptr,4,%g0 ! res_ptr unaligned? Side effect: cy=0 - be L_v1b ! if no, branch + andcc RES_PTR,4,%g0 ! RES_PTR unaligned? Side effect: cy=0 + be LOC(v1b) ! if no, branch nop -/* Add least significant limb separately to align res_ptr and s1_ptr */ - ld [s2_ptr],%g4 - add s2_ptr,4,s2_ptr - ld [s1_ptr],%g2 - add s1_ptr,4,s1_ptr - add size,-1,size +/* Add least significant limb separately to align RES_PTR and S1_PTR */ + ld [S2_PTR],%g4 + add S2_PTR,4,S2_PTR + ld [S1_PTR],%g2 + add S1_PTR,4,S1_PTR + add SIZE,-1,SIZE subcc %g2,%g4,%o4 - st %o4,[res_ptr] - add res_ptr,4,res_ptr -L_v1b: addx %g0,%g0,%o4 ! save cy in register - cmp size,2 ! if size < 2 ... - bl Lend2 ! ... branch to tail code + st %o4,[RES_PTR] + add RES_PTR,4,RES_PTR +LOC(v1b): + addx %g0,%g0,%o4 ! save cy in register + cmp SIZE,2 ! if SIZE < 2 ... + bl LOC(end2) ! ... branch to tail code subcc %g0,%o4,%g0 ! restore cy - ld [s2_ptr+0],%g4 - addcc size,-10,size - ld [s2_ptr+4],%g1 - ldd [s1_ptr+0],%g2 - blt Lfin1b + ld [S2_PTR+0],%g4 + addcc SIZE,-10,SIZE + ld [S2_PTR+4],%g1 + ldd [S1_PTR+0],%g2 + blt LOC(fin1b) subcc %g0,%o4,%g0 ! restore cy /* Add blocks of 8 limbs until less than 8 limbs remain */ -Loop1b: subxcc %g2,%g4,%o4 - ld [s2_ptr+8],%g4 +LOC(loop1b): + subxcc %g2,%g4,%o4 + ld [S2_PTR+8],%g4 subxcc %g3,%g1,%o5 - ld [s2_ptr+12],%g1 - ldd [s1_ptr+8],%g2 - std %o4,[res_ptr+0] + ld [S2_PTR+12],%g1 + ldd [S1_PTR+8],%g2 + std %o4,[RES_PTR+0] subxcc %g2,%g4,%o4 - ld [s2_ptr+16],%g4 + ld [S2_PTR+16],%g4 subxcc %g3,%g1,%o5 - ld [s2_ptr+20],%g1 - ldd [s1_ptr+16],%g2 - std %o4,[res_ptr+8] + ld [S2_PTR+20],%g1 + ldd [S1_PTR+16],%g2 + std %o4,[RES_PTR+8] subxcc %g2,%g4,%o4 - ld [s2_ptr+24],%g4 + ld [S2_PTR+24],%g4 subxcc %g3,%g1,%o5 - ld [s2_ptr+28],%g1 - ldd [s1_ptr+24],%g2 - std %o4,[res_ptr+16] + ld [S2_PTR+28],%g1 + ldd [S1_PTR+24],%g2 + std %o4,[RES_PTR+16] subxcc %g2,%g4,%o4 - ld [s2_ptr+32],%g4 + ld [S2_PTR+32],%g4 subxcc %g3,%g1,%o5 - ld [s2_ptr+36],%g1 - ldd [s1_ptr+32],%g2 - std %o4,[res_ptr+24] + ld [S2_PTR+36],%g1 + ldd [S1_PTR+32],%g2 + std %o4,[RES_PTR+24] addx %g0,%g0,%o4 ! save cy in register - addcc size,-8,size - add s1_ptr,32,s1_ptr - add s2_ptr,32,s2_ptr - add res_ptr,32,res_ptr - bge Loop1b + addcc SIZE,-8,SIZE + add S1_PTR,32,S1_PTR + add S2_PTR,32,S2_PTR + add RES_PTR,32,RES_PTR + bge LOC(loop1b) subcc %g0,%o4,%g0 ! restore cy -Lfin1b: addcc size,8-2,size - blt Lend1b +LOC(fin1b): + addcc SIZE,8-2,SIZE + blt LOC(end1b) subcc %g0,%o4,%g0 ! restore cy /* Add blocks of 2 limbs until less than 2 limbs remain */ -Loope1b:subxcc %g2,%g4,%o4 - ld [s2_ptr+8],%g4 +LOC(loope1b): + subxcc %g2,%g4,%o4 + ld [S2_PTR+8],%g4 subxcc %g3,%g1,%o5 - ld [s2_ptr+12],%g1 - ldd [s1_ptr+8],%g2 - std %o4,[res_ptr+0] + ld [S2_PTR+12],%g1 + ldd [S1_PTR+8],%g2 + std %o4,[RES_PTR+0] addx %g0,%g0,%o4 ! save cy in register - addcc size,-2,size - add s1_ptr,8,s1_ptr - add s2_ptr,8,s2_ptr - add res_ptr,8,res_ptr - bge Loope1b + addcc SIZE,-2,SIZE + add S1_PTR,8,S1_PTR + add S2_PTR,8,S2_PTR + add RES_PTR,8,RES_PTR + bge LOC(loope1b) subcc %g0,%o4,%g0 ! restore cy -Lend1b: subxcc %g2,%g4,%o4 +LOC(end1b): + subxcc %g2,%g4,%o4 subxcc %g3,%g1,%o5 - std %o4,[res_ptr+0] + std %o4,[RES_PTR+0] addx %g0,%g0,%o4 ! save cy in register - andcc size,1,%g0 - be Lret1b + andcc SIZE,1,%g0 + be LOC(ret1b) subcc %g0,%o4,%g0 ! restore cy /* Add last limb */ - ld [s2_ptr+8],%g4 - ld [s1_ptr+8],%g2 + ld [S2_PTR+8],%g4 + ld [S1_PTR+8],%g2 subxcc %g2,%g4,%o4 - st %o4,[res_ptr+8] + st %o4,[RES_PTR+8] -Lret1b: retl +LOC(ret1b): + retl addx %g0,%g0,%o0 ! return carry-out from most sign. limb ! ** V2 ** -/* If we come here, the alignment of s1_ptr and res_ptr as well as the - alignment of s2_ptr and res_ptr differ. Since there are only two ways +/* If we come here, the alignment of S1_PTR and RES_PTR as well as the + alignment of S2_PTR and RES_PTR differ. Since there are only two ways things can be aligned (that we care about) we now know that the alignment - of s1_ptr and s2_ptr are the same. */ + of S1_PTR and S2_PTR are the same. */ -L2: cmp size,1 - be Ljone +LOC(2): cmp SIZE,1 + be LOC(jone) nop - andcc s1_ptr,4,%g0 ! s1_ptr unaligned? Side effect: cy=0 - be L_v2 ! if no, branch + andcc S1_PTR,4,%g0 ! S1_PTR unaligned? Side effect: cy=0 + be LOC(v2) ! if no, branch nop -/* Add least significant limb separately to align s1_ptr and s2_ptr */ - ld [s1_ptr],%g4 - add s1_ptr,4,s1_ptr - ld [s2_ptr],%g2 - add s2_ptr,4,s2_ptr - add size,-1,size +/* Add least significant limb separately to align S1_PTR and S2_PTR */ + ld [S1_PTR],%g4 + add S1_PTR,4,S1_PTR + ld [S2_PTR],%g2 + add S2_PTR,4,S2_PTR + add SIZE,-1,SIZE subcc %g4,%g2,%o4 - st %o4,[res_ptr] - add res_ptr,4,res_ptr + st %o4,[RES_PTR] + add RES_PTR,4,RES_PTR -L_v2: addx %g0,%g0,%o4 ! save cy in register - addcc size,-8,size - blt Lfin2 +LOC(v2): + addx %g0,%g0,%o4 ! save cy in register + addcc SIZE,-8,SIZE + blt LOC(fin2) subcc %g0,%o4,%g0 ! restore cy /* Add blocks of 8 limbs until less than 8 limbs remain */ -Loop2: ldd [s1_ptr+0],%g2 - ldd [s2_ptr+0],%o4 +LOC(loop2): + ldd [S1_PTR+0],%g2 + ldd [S2_PTR+0],%o4 subxcc %g2,%o4,%g2 - st %g2,[res_ptr+0] + st %g2,[RES_PTR+0] subxcc %g3,%o5,%g3 - st %g3,[res_ptr+4] - ldd [s1_ptr+8],%g2 - ldd [s2_ptr+8],%o4 + st %g3,[RES_PTR+4] + ldd [S1_PTR+8],%g2 + ldd [S2_PTR+8],%o4 subxcc %g2,%o4,%g2 - st %g2,[res_ptr+8] + st %g2,[RES_PTR+8] subxcc %g3,%o5,%g3 - st %g3,[res_ptr+12] - ldd [s1_ptr+16],%g2 - ldd [s2_ptr+16],%o4 + st %g3,[RES_PTR+12] + ldd [S1_PTR+16],%g2 + ldd [S2_PTR+16],%o4 subxcc %g2,%o4,%g2 - st %g2,[res_ptr+16] + st %g2,[RES_PTR+16] subxcc %g3,%o5,%g3 - st %g3,[res_ptr+20] - ldd [s1_ptr+24],%g2 - ldd [s2_ptr+24],%o4 + st %g3,[RES_PTR+20] + ldd [S1_PTR+24],%g2 + ldd [S2_PTR+24],%o4 subxcc %g2,%o4,%g2 - st %g2,[res_ptr+24] + st %g2,[RES_PTR+24] subxcc %g3,%o5,%g3 - st %g3,[res_ptr+28] + st %g3,[RES_PTR+28] addx %g0,%g0,%o4 ! save cy in register - addcc size,-8,size - add s1_ptr,32,s1_ptr - add s2_ptr,32,s2_ptr - add res_ptr,32,res_ptr - bge Loop2 + addcc SIZE,-8,SIZE + add S1_PTR,32,S1_PTR + add S2_PTR,32,S2_PTR + add RES_PTR,32,RES_PTR + bge LOC(loop2) subcc %g0,%o4,%g0 ! restore cy -Lfin2: addcc size,8-2,size - blt Lend2 +LOC(fin2): + addcc SIZE,8-2,SIZE + blt LOC(end2) subcc %g0,%o4,%g0 ! restore cy -Loope2: ldd [s1_ptr+0],%g2 - ldd [s2_ptr+0],%o4 +LOC(loope2): + ldd [S1_PTR+0],%g2 + ldd [S2_PTR+0],%o4 subxcc %g2,%o4,%g2 - st %g2,[res_ptr+0] + st %g2,[RES_PTR+0] subxcc %g3,%o5,%g3 - st %g3,[res_ptr+4] + st %g3,[RES_PTR+4] addx %g0,%g0,%o4 ! save cy in register - addcc size,-2,size - add s1_ptr,8,s1_ptr - add s2_ptr,8,s2_ptr - add res_ptr,8,res_ptr - bge Loope2 + addcc SIZE,-2,SIZE + add S1_PTR,8,S1_PTR + add S2_PTR,8,S2_PTR + add RES_PTR,8,RES_PTR + bge LOC(loope2) subcc %g0,%o4,%g0 ! restore cy -Lend2: andcc size,1,%g0 - be Lret2 +LOC(end2): + andcc SIZE,1,%g0 + be LOC(ret2) subcc %g0,%o4,%g0 ! restore cy /* Add last limb */ -Ljone: ld [s1_ptr],%g4 - ld [s2_ptr],%g2 +LOC(jone): + ld [S1_PTR],%g4 + ld [S2_PTR],%g2 subxcc %g4,%g2,%o4 - st %o4,[res_ptr] + st %o4,[RES_PTR] -Lret2: retl +LOC(ret2): + retl addx %g0,%g0,%o0 ! return carry-out from most sign. limb + +END(__mpn_sub_n) |