diff options
author | Zack Weinberg <zackw@panix.com> | 2017-06-08 15:39:03 -0400 |
---|---|---|
committer | Zack Weinberg <zackw@panix.com> | 2017-06-08 15:39:03 -0400 |
commit | 5046dbb4a7eba5eccfd258f92f4735c9ffc8d069 (patch) | |
tree | 4470480d904b65cf14ca524f96f79eca818c3eaf /REORG.TODO/sysdeps/i386/i586/lshift.S | |
parent | 199fc19d3aaaf57944ef036e15904febe877fc93 (diff) | |
download | glibc-zack/build-layout-experiment.tar glibc-zack/build-layout-experiment.tar.gz glibc-zack/build-layout-experiment.tar.bz2 glibc-zack/build-layout-experiment.zip |
Prepare for radical source tree reorganization.zack/build-layout-experiment
All top-level files and directories are moved into a temporary storage
directory, REORG.TODO, except for files that will certainly still
exist in their current form at top level when we're done (COPYING,
COPYING.LIB, LICENSES, NEWS, README), all old ChangeLog files (which
are moved to the new directory OldChangeLogs, instead), and the
generated file INSTALL (which is just deleted; in the new order, there
will be no generated files checked into version control).
Diffstat (limited to 'REORG.TODO/sysdeps/i386/i586/lshift.S')
-rw-r--r-- | REORG.TODO/sysdeps/i386/i586/lshift.S | 255 |
1 files changed, 255 insertions, 0 deletions
diff --git a/REORG.TODO/sysdeps/i386/i586/lshift.S b/REORG.TODO/sysdeps/i386/i586/lshift.S new file mode 100644 index 0000000000..7941c28d9d --- /dev/null +++ b/REORG.TODO/sysdeps/i386/i586/lshift.S @@ -0,0 +1,255 @@ +/* Pentium optimized __mpn_lshift -- + Copyright (C) 1992-2017 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + <http://www.gnu.org/licenses/>. */ + +#include "sysdep.h" +#include "asm-syntax.h" + +#define PARMS 4+16 /* space for 4 saved regs */ +#define RES PARMS +#define S RES+4 +#define SIZE S+4 +#define CNT SIZE+4 + + .text +ENTRY (__mpn_lshift) + + pushl %edi + cfi_adjust_cfa_offset (4) + pushl %esi + cfi_adjust_cfa_offset (4) + pushl %ebp + cfi_adjust_cfa_offset (4) + cfi_rel_offset (ebp, 0) + pushl %ebx + cfi_adjust_cfa_offset (4) + + movl RES(%esp),%edi + cfi_rel_offset (edi, 12) + movl S(%esp),%esi + cfi_rel_offset (esi, 8) + movl SIZE(%esp),%ebx + cfi_rel_offset (ebx, 0) + movl CNT(%esp),%ecx + +/* We can use faster code for shift-by-1 under certain conditions. */ + cmp $1,%ecx + jne L(normal) + leal 4(%esi),%eax + cmpl %edi,%eax + jnc L(special) /* jump if s_ptr + 1 >= res_ptr */ + leal (%esi,%ebx,4),%eax + cmpl %eax,%edi + jnc L(special) /* jump if res_ptr >= s_ptr + size */ + +L(normal): + leal -4(%edi,%ebx,4),%edi + leal -4(%esi,%ebx,4),%esi + + movl (%esi),%edx + subl $4,%esi + xorl %eax,%eax + shldl %cl,%edx,%eax /* compute carry limb */ + pushl %eax /* push carry limb onto stack */ + cfi_adjust_cfa_offset (4) + + decl %ebx + pushl %ebx + cfi_adjust_cfa_offset (4) + shrl $3,%ebx + jz L(end) + + movl (%edi),%eax /* fetch destination cache line */ + + ALIGN (2) +L(oop): movl -28(%edi),%eax /* fetch destination cache line */ + movl %edx,%ebp + + movl (%esi),%eax + movl -4(%esi),%edx + shldl %cl,%eax,%ebp + shldl %cl,%edx,%eax + movl %ebp,(%edi) + movl %eax,-4(%edi) + + movl -8(%esi),%ebp + movl -12(%esi),%eax + shldl %cl,%ebp,%edx + shldl %cl,%eax,%ebp + movl %edx,-8(%edi) + movl %ebp,-12(%edi) + + movl -16(%esi),%edx + movl -20(%esi),%ebp + shldl %cl,%edx,%eax + shldl %cl,%ebp,%edx + movl %eax,-16(%edi) + movl %edx,-20(%edi) + + movl -24(%esi),%eax + movl -28(%esi),%edx + shldl %cl,%eax,%ebp + shldl %cl,%edx,%eax + movl %ebp,-24(%edi) + movl %eax,-28(%edi) + + subl $32,%esi + subl $32,%edi + decl %ebx + jnz L(oop) + +L(end): popl %ebx + cfi_adjust_cfa_offset (-4) + andl $7,%ebx + jz L(end2) +L(oop2): + movl (%esi),%eax + shldl %cl,%eax,%edx + movl %edx,(%edi) + movl %eax,%edx + subl $4,%esi + subl $4,%edi + decl %ebx + jnz L(oop2) + +L(end2): + shll %cl,%edx /* compute least significant limb */ + movl %edx,(%edi) /* store it */ + + popl %eax /* pop carry limb */ + cfi_adjust_cfa_offset (-4) + + popl %ebx + cfi_adjust_cfa_offset (-4) + cfi_restore (ebx) + popl %ebp + cfi_adjust_cfa_offset (-4) + cfi_restore (ebp) + popl %esi + cfi_adjust_cfa_offset (-4) + cfi_restore (esi) + popl %edi + cfi_adjust_cfa_offset (-4) + cfi_restore (edi) + + ret + +/* We loop from least significant end of the arrays, which is only + permissible if the source and destination don't overlap, since the + function is documented to work for overlapping source and destination. +*/ + + cfi_adjust_cfa_offset (16) + cfi_rel_offset (edi, 12) + cfi_rel_offset (esi, 8) + cfi_rel_offset (ebp, 4) + cfi_rel_offset (ebx, 0) +L(special): + movl (%esi),%edx + addl $4,%esi + + decl %ebx + pushl %ebx + cfi_adjust_cfa_offset (4) + shrl $3,%ebx + + addl %edx,%edx + incl %ebx + decl %ebx + jz L(Lend) + + movl (%edi),%eax /* fetch destination cache line */ + + ALIGN (2) +L(Loop): + movl 28(%edi),%eax /* fetch destination cache line */ + movl %edx,%ebp + + movl (%esi),%eax + movl 4(%esi),%edx + adcl %eax,%eax + movl %ebp,(%edi) + adcl %edx,%edx + movl %eax,4(%edi) + + movl 8(%esi),%ebp + movl 12(%esi),%eax + adcl %ebp,%ebp + movl %edx,8(%edi) + adcl %eax,%eax + movl %ebp,12(%edi) + + movl 16(%esi),%edx + movl 20(%esi),%ebp + adcl %edx,%edx + movl %eax,16(%edi) + adcl %ebp,%ebp + movl %edx,20(%edi) + + movl 24(%esi),%eax + movl 28(%esi),%edx + adcl %eax,%eax + movl %ebp,24(%edi) + adcl %edx,%edx + movl %eax,28(%edi) + + leal 32(%esi),%esi /* use leal not to clobber carry */ + leal 32(%edi),%edi + decl %ebx + jnz L(Loop) + +L(Lend): + popl %ebx + cfi_adjust_cfa_offset (-4) + sbbl %eax,%eax /* save carry in %eax */ + andl $7,%ebx + jz L(Lend2) + addl %eax,%eax /* restore carry from eax */ +L(Loop2): + movl %edx,%ebp + movl (%esi),%edx + adcl %edx,%edx + movl %ebp,(%edi) + + leal 4(%esi),%esi /* use leal not to clobber carry */ + leal 4(%edi),%edi + decl %ebx + jnz L(Loop2) + + jmp L(L1) +L(Lend2): + addl %eax,%eax /* restore carry from eax */ +L(L1): movl %edx,(%edi) /* store last limb */ + + sbbl %eax,%eax + negl %eax + + popl %ebx + cfi_adjust_cfa_offset (-4) + cfi_restore (ebx) + popl %ebp + cfi_adjust_cfa_offset (-4) + cfi_restore (ebp) + popl %esi + cfi_adjust_cfa_offset (-4) + cfi_restore (esi) + popl %edi + cfi_adjust_cfa_offset (-4) + cfi_restore (edi) + + ret +END (__mpn_lshift) |