aboutsummaryrefslogtreecommitdiff
path: root/REORG.TODO/sysdeps/x86_64/add_n.S
diff options
context:
space:
mode:
Diffstat (limited to 'REORG.TODO/sysdeps/x86_64/add_n.S')
-rw-r--r--REORG.TODO/sysdeps/x86_64/add_n.S100
1 files changed, 100 insertions, 0 deletions
diff --git a/REORG.TODO/sysdeps/x86_64/add_n.S b/REORG.TODO/sysdeps/x86_64/add_n.S
new file mode 100644
index 0000000000..4ba83c0bdb
--- /dev/null
+++ b/REORG.TODO/sysdeps/x86_64/add_n.S
@@ -0,0 +1,100 @@
+/* x86-64 __mpn_add_n -- Add two limb vectors of the same length > 0 and store
+ sum in a third limb vector.
+ Copyright (C) 2006-2017 Free Software Foundation, Inc.
+ This file is part of the GNU MP Library.
+
+ The GNU MP Library is free software; you can redistribute it and/or modify
+ it under the terms of the GNU Lesser General Public License as published by
+ the Free Software Foundation; either version 2.1 of the License, or (at your
+ option) any later version.
+
+ The GNU MP Library is distributed in the hope that it will be useful, but
+ WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
+ or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
+ License for more details.
+
+ You should have received a copy of the GNU Lesser General Public License
+ along with the GNU MP Library; see the file COPYING.LIB. If not,
+ see <http://www.gnu.org/licenses/>. */
+
+#include "sysdep.h"
+#include "asm-syntax.h"
+
+#define rp %rdi
+#define up %rsi
+#define vp %rdx
+#define n %rcx
+#define cy %r8
+
+#ifndef func
+# define func __mpn_add_n
+# define ADCSBB adc
+#endif
+
+ .text
+ENTRY (func)
+ xor %r8, %r8
+ mov (up), %r10
+ mov (vp), %r11
+
+ lea -8(up,n,8), up
+ lea -8(vp,n,8), vp
+ lea -16(rp,n,8), rp
+ mov %ecx, %eax
+ neg n
+ and $3, %eax
+ je L(b00)
+ add %rax, n /* clear low rcx bits for jrcxz */
+ cmp $2, %eax
+ jl L(b01)
+ je L(b10)
+
+L(b11): shr %r8 /* set cy */
+ jmp L(e11)
+
+L(b00): shr %r8 /* set cy */
+ mov %r10, %r8
+ mov %r11, %r9
+ lea 4(n), n
+ jmp L(e00)
+
+L(b01): shr %r8 /* set cy */
+ jmp L(e01)
+
+L(b10): shr %r8 /* set cy */
+ mov %r10, %r8
+ mov %r11, %r9
+ jmp L(e10)
+
+L(end): ADCSBB %r11, %r10
+ mov %r10, 8(rp)
+ mov %ecx, %eax /* clear eax, ecx contains 0 */
+ adc %eax, %eax
+ ret
+
+ .p2align 4
+L(top):
+ mov -24(up,n,8), %r8
+ mov -24(vp,n,8), %r9
+ ADCSBB %r11, %r10
+ mov %r10, -24(rp,n,8)
+L(e00):
+ mov -16(up,n,8), %r10
+ mov -16(vp,n,8), %r11
+ ADCSBB %r9, %r8
+ mov %r8, -16(rp,n,8)
+L(e11):
+ mov -8(up,n,8), %r8
+ mov -8(vp,n,8), %r9
+ ADCSBB %r11, %r10
+ mov %r10, -8(rp,n,8)
+L(e10):
+ mov (up,n,8), %r10
+ mov (vp,n,8), %r11
+ ADCSBB %r9, %r8
+ mov %r8, (rp,n,8)
+L(e01):
+ jrcxz L(end)
+ lea 4(n), n
+ jmp L(top)
+END (func)