aboutsummaryrefslogtreecommitdiff
path: root/sysdeps/powerpc/powerpc64/power6
diff options
context:
space:
mode:
authorZack Weinberg <zackw@panix.com>2017-06-08 15:39:03 -0400
committerZack Weinberg <zackw@panix.com>2017-06-08 15:39:03 -0400
commit5046dbb4a7eba5eccfd258f92f4735c9ffc8d069 (patch)
tree4470480d904b65cf14ca524f96f79eca818c3eaf /sysdeps/powerpc/powerpc64/power6
parent199fc19d3aaaf57944ef036e15904febe877fc93 (diff)
downloadglibc-zack/build-layout-experiment.tar
glibc-zack/build-layout-experiment.tar.gz
glibc-zack/build-layout-experiment.tar.bz2
glibc-zack/build-layout-experiment.zip
Prepare for radical source tree reorganization.zack/build-layout-experiment
All top-level files and directories are moved into a temporary storage directory, REORG.TODO, except for files that will certainly still exist in their current form at top level when we're done (COPYING, COPYING.LIB, LICENSES, NEWS, README), all old ChangeLog files (which are moved to the new directory OldChangeLogs, instead), and the generated file INSTALL (which is just deleted; in the new order, there will be no generated files checked into version control).
Diffstat (limited to 'sysdeps/powerpc/powerpc64/power6')
-rw-r--r--sysdeps/powerpc/powerpc64/power6/Implies2
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/Implies1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S58
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S59
-rw-r--r--sysdeps/powerpc/powerpc64/power6/memcpy.S1499
-rw-r--r--sysdeps/powerpc/powerpc64/power6/memset.S395
-rw-r--r--sysdeps/powerpc/powerpc64/power6/multiarch/Implies1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/wcschr.c1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/wcscpy.c1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/wcsrchr.c1
12 files changed, 0 insertions, 2020 deletions
diff --git a/sysdeps/powerpc/powerpc64/power6/Implies b/sysdeps/powerpc/powerpc64/power6/Implies
deleted file mode 100644
index 4c782d4122..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/Implies
+++ /dev/null
@@ -1,2 +0,0 @@
-powerpc/powerpc64/power5+/fpu
-powerpc/powerpc64/power5+
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/Implies b/sysdeps/powerpc/powerpc64/power6/fpu/Implies
deleted file mode 100644
index f09854edb6..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/Implies
+++ /dev/null
@@ -1 +0,0 @@
-powerpc/powerpc64/power5+/fpu
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies b/sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies
deleted file mode 100644
index fca8a4ef0f..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies
+++ /dev/null
@@ -1 +0,0 @@
-powerpc/powerpc64/power5+/fpu/multiarch
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S b/sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S
deleted file mode 100644
index ec36d1be5b..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S
+++ /dev/null
@@ -1,58 +0,0 @@
-/* copysign(). PowerPC64/POWER6 version.
- Copyright (C) 2010-2017 Free Software Foundation, Inc.
- Contributed by Luis Machado <luisgpm@br.ibm.com>.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-#include <math_ldbl_opt.h>
-
-/* double [f1] copysign (double [f1] x, double [f2] y);
- copysign(x,y) returns a value with the magnitude of x and
- with the sign bit of y. */
-
- .section ".text"
- .type __copysign, @function
- .machine power6
-EALIGN (__copysign, 4, 0)
- CALL_MCOUNT 0
- fcpsgn fp1,fp2,fp1
- blr
-END (__copysign)
-
-hidden_def (__copysign)
-weak_alias (__copysign, copysign)
-
-/* It turns out that the 'double' version will also always work for
- single-precision. */
-strong_alias (__copysign, __copysignf)
-hidden_def (__copysignf)
-weak_alias (__copysignf, copysignf)
-
-#ifdef NO_LONG_DOUBLE
-strong_alias (__copysign, __copysignl)
-weak_alias (__copysign, copysignl)
-#endif
-
-#if IS_IN (libm)
-# if LONG_DOUBLE_COMPAT (libm, GLIBC_2_0)
-compat_symbol (libm, copysign, copysignl, GLIBC_2_0)
-# endif
-#else
-# if LONG_DOUBLE_COMPAT(libc, GLIBC_2_0)
-compat_symbol (libc, copysign, copysignl, GLIBC_2_0);
-# endif
-#endif
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S b/sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S
deleted file mode 100644
index d4aa702d07..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S
+++ /dev/null
@@ -1 +0,0 @@
-/* This function uses the same code as s_copysign.S. */
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S b/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
deleted file mode 100644
index 85187b45f3..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
+++ /dev/null
@@ -1,59 +0,0 @@
-/* isnan(). PowerPC64 version.
- Copyright (C) 2008-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-#include <math_ldbl_opt.h>
-
-/* int __isnan(x) */
- .machine power6
-EALIGN (__isnan, 4, 0)
- CALL_MCOUNT 0
- stfd fp1,-8(r1) /* copy FPR to GPR */
- ori r1,r1,0
- ld r4,-8(r1)
- lis r0,0x7ff0
- sldi r0,r0,32 /* const long r0 0x7ff00000 00000000 */
- clrldi r4,r4,1 /* x = fabs(x) */
- cmpd cr7,r4,r0 /* if (fabs(x) <= inf) */
- li r3,0 /* then return 0 */
- blelr+ cr7
- li r3,1 /* else return 1 */
- blr
- END (__isnan)
-
-hidden_def (__isnan)
-weak_alias (__isnan, isnan)
-
-/* It turns out that the 'double' version will also always work for
- single-precision. */
-strong_alias (__isnan, __isnanf)
-hidden_def (__isnanf)
-weak_alias (__isnanf, isnanf)
-
-#ifdef NO_LONG_DOUBLE
-strong_alias (__isnan, __isnanl)
-weak_alias (__isnan, isnanl)
-#endif
-
-#if !IS_IN (libm)
-# if LONG_DOUBLE_COMPAT(libc, GLIBC_2_0)
-compat_symbol (libc, __isnan, __isnanl, GLIBC_2_0);
-compat_symbol (libc, isnan, isnanl, GLIBC_2_0);
-# endif
-#endif
-
diff --git a/sysdeps/powerpc/powerpc64/power6/memcpy.S b/sysdeps/powerpc/powerpc64/power6/memcpy.S
deleted file mode 100644
index 1f7294b8ed..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/memcpy.S
+++ /dev/null
@@ -1,1499 +0,0 @@
-/* Optimized memcpy implementation for PowerPC64.
- Copyright (C) 2003-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-
-/* __ptr_t [r3] memcpy (__ptr_t dst [r3], __ptr_t src [r4], size_t len [r5]);
- Returns 'dst'.
-
- Memcpy handles short copies (< 32-bytes) using a binary move blocks
- (no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
- with the appropriate combination of byte and halfword load/stores.
- There is minimal effort to optimize the alignment of short moves.
- The 64-bit implementations of POWER3 and POWER4 do a reasonable job
- of handling unaligned load/stores that do not cross 32-byte boundaries.
-
- Longer moves (>= 32-bytes) justify the effort to get at least the
- destination doubleword (8-byte) aligned. Further optimization is
- possible when both source and destination are doubleword aligned.
- Each case has a optimized unrolled loop.
-
- For POWER6 unaligned loads will take a 20+ cycle hiccup for any
- L1 cache miss that crosses a 32- or 128-byte boundary. Store
- is more forgiving and does not take a hiccup until page or
- segment boundaries. So we require doubleword alignment for
- the source but may take a risk and only require word alignment
- for the destination. */
-
-#ifndef MEMCPY
-# define MEMCPY memcpy
-#endif
- .machine "power6"
-EALIGN (MEMCPY, 7, 0)
- CALL_MCOUNT 3
-
- cmpldi cr1,5,31
- neg 0,3
- std 3,-16(1)
- std 31,-8(1)
- andi. 11,3,7 /* check alignment of dst. */
- clrldi 0,0,61 /* Number of bytes until the 1st doubleword of dst. */
- clrldi 10,4,61 /* check alignment of src. */
- cmpldi cr6,5,8
- ble- cr1,.L2 /* If move < 32 bytes use short move code. */
- mtcrf 0x01,0
- cmpld cr6,10,11
- srdi 9,5,3 /* Number of full double words remaining. */
- beq .L0
-
- subf 5,0,5
- /* Move 0-7 bytes as needed to get the destination doubleword aligned.
- Duplicate some code to maximize fall-through and minimize agen delays. */
-1: bf 31,2f
- lbz 6,0(4)
- stb 6,0(3)
- bf 30,5f
- lhz 6,1(4)
- sth 6,1(3)
- bf 29,0f
- lwz 6,3(4)
- stw 6,3(3)
- b 0f
-5:
- bf 29,0f
- lwz 6,1(4)
- stw 6,1(3)
- b 0f
-
-2: bf 30,4f
- lhz 6,0(4)
- sth 6,0(3)
- bf 29,0f
- lwz 6,2(4)
- stw 6,2(3)
- b 0f
-
-4: bf 29,0f
- lwz 6,0(4)
- stw 6,0(3)
-0:
-/* Add the number of bytes until the 1st doubleword of dst to src and dst. */
- add 4,4,0
- add 3,3,0
-
- clrldi 10,4,61 /* check alignment of src again. */
- srdi 9,5,3 /* Number of full double words remaining. */
-
- /* Copy doublewords from source to destination, assuming the
- destination is aligned on a doubleword boundary.
-
- At this point we know there are at least 25 bytes left (32-7) to copy.
- The next step is to determine if the source is also doubleword aligned.
- If not branch to the unaligned move code at .L6. which uses
- a load, shift, store strategy.
-
- Otherwise source and destination are doubleword aligned, and we can
- the optimized doubleword copy loop. */
- .align 4
-.L0:
- clrldi 11,5,61
- andi. 0,5,0x78
- srdi 12,5,7 /* Number of 128-byte blocks to move. */
- cmpldi cr1,11,0 /* If the tail is 0 bytes */
- bne- cr6,.L6 /* If source is not DW aligned. */
-
- /* Move doublewords where destination and source are DW aligned.
- Use a unrolled loop to copy 16 doublewords (128-bytes) per iteration.
- If the copy is not an exact multiple of 128 bytes, 1-15
- doublewords are copied as needed to set up the main loop. After
- the main loop exits there may be a tail of 1-7 bytes. These byte
- are copied a word/halfword/byte at a time as needed to preserve
- alignment.
-
- For POWER6 the L1 is store-through and the L2 is store-in. The
- L2 is clocked at half CPU clock so we can store 16 bytes every
- other cycle. POWER6 also has a load/store bypass so we can do
- load, load, store, store every 2 cycles.
-
- The following code is sensitive to cache line alignment. Do not
- make any change with out first making sure they don't result in
- splitting ld/std pairs across a cache line. */
-
- mtcrf 0x02,5
- mtcrf 0x01,5
- cmpldi cr5,12,1
- beq L(das_loop)
-
- bf 25,4f
- .align 3
- ld 6,0(4)
- ld 7,8(4)
- mr 11,4
- mr 10,3
- std 6,0(3)
- std 7,8(3)
- ld 6,16(4)
- ld 7,24(4)
- std 6,16(3)
- std 7,24(3)
- ld 6,0+32(4)
- ld 7,8+32(4)
- addi 4,4,64
- addi 3,3,64
- std 6,0+32(10)
- std 7,8+32(10)
- ld 6,16+32(11)
- ld 7,24+32(11)
- std 6,16+32(10)
- std 7,24+32(10)
-4:
- mr 10,3
- bf 26,2f
- ld 6,0(4)
- ld 7,8(4)
- mr 11,4
- nop
- std 6,0(3)
- std 7,8(3)
- ld 6,16(4)
- ld 7,24(4)
- addi 4,4,32
- std 6,16(3)
- std 7,24(3)
- addi 3,3,32
-6:
- nop
- bf 27,5f
- ld 6,0+32(11)
- ld 7,8+32(11)
- addi 4,4,16
- addi 3,3,16
- std 6,0+32(10)
- std 7,8+32(10)
- bf 28,L(das_loop_s)
- ld 0,16+32(11)
- addi 4,4,8
- addi 3,3,8
- std 0,16+32(10)
- blt cr5,L(das_tail)
- b L(das_loop)
- .align 3
-5:
- nop
- bf 28,L(das_loop_s)
- ld 6,32(11)
- addi 4,4,8
- addi 3,3,8
- std 6,32(10)
- blt cr5,L(das_tail)
- b L(das_loop)
- .align 3
-2:
- mr 11,4
- bf 27,1f
- ld 6,0(4)
- ld 7,8(4)
- addi 4,4,16
- addi 3,3,16
- std 6,0(10)
- std 7,8(10)
- bf 28,L(das_loop_s)
- ld 0,16(11)
- addi 4,11,24
- addi 3,10,24
- std 0,16(10)
- blt cr5,L(das_tail)
- b L(das_loop)
- .align 3
-1:
- nop
- bf 28,L(das_loop_s)
- ld 6,0(4)
- addi 4,4,8
- addi 3,3,8
- std 6,0(10)
-L(das_loop_s):
- nop
- blt cr5,L(das_tail)
- .align 4
-L(das_loop):
- ld 6,0(4)
- ld 7,8(4)
- mr 10,3
- mr 11,4
- std 6,0(3)
- std 7,8(3)
- addi 12,12,-1
- nop
- ld 8,16(4)
- ld 0,24(4)
- std 8,16(3)
- std 0,24(3)
-
- ld 6,0+32(4)
- ld 7,8+32(4)
- std 6,0+32(3)
- std 7,8+32(3)
- ld 8,16+32(4)
- ld 0,24+32(4)
- std 8,16+32(3)
- std 0,24+32(3)
-
- ld 6,0+64(11)
- ld 7,8+64(11)
- std 6,0+64(10)
- std 7,8+64(10)
- ld 8,16+64(11)
- ld 0,24+64(11)
- std 8,16+64(10)
- std 0,24+64(10)
-
- ld 6,0+96(11)
- ld 7,8+96(11)
- addi 4,4,128
- addi 3,3,128
- std 6,0+96(10)
- std 7,8+96(10)
- ld 8,16+96(11)
- ld 0,24+96(11)
- std 8,16+96(10)
- std 0,24+96(10)
- ble cr5,L(das_loop_e)
-
- mtctr 12
- .align 4
-L(das_loop2):
- ld 6,0(4)
- ld 7,8(4)
- mr 10,3
- mr 11,4
- std 6,0(3)
- std 7,8(3)
- ld 8,16(4)
- ld 0,24(4)
- std 8,16(3)
- std 0,24(3)
-
- ld 6,0+32(4)
- ld 7,8+32(4)
- std 6,0+32(3)
- std 7,8+32(3)
- ld 8,16+32(4)
- ld 0,24+32(4)
- std 8,16+32(3)
- std 0,24+32(3)
-
- ld 6,0+64(11)
- ld 7,8+64(11)
- std 6,0+64(10)
- std 7,8+64(10)
- ld 8,16+64(11)
- ld 0,24+64(11)
- std 8,16+64(10)
- std 0,24+64(10)
-
- ld 6,0+96(11)
- ld 7,8+96(11)
- addi 4,4,128
- addi 3,3,128
- std 6,0+96(10)
- std 7,8+96(10)
- ld 8,16+96(11)
- ld 0,24+96(11)
- std 8,16+96(10)
- std 0,24+96(10)
- bdnz L(das_loop2)
-L(das_loop_e):
-/* Check of a 1-7 byte tail, return if none. */
- bne cr1,L(das_tail2)
-/* Return original dst pointer. */
- ld 3,-16(1)
- blr
- .align 4
-L(das_tail):
- beq cr1,0f
-
-L(das_tail2):
-/* At this point we have a tail of 0-7 bytes and we know that the
- destination is double word aligned. */
-4: bf 29,2f
- lwz 6,0(4)
- stw 6,0(3)
- bf 30,5f
- lhz 6,4(4)
- sth 6,4(3)
- bf 31,0f
- lbz 6,6(4)
- stb 6,6(3)
- b 0f
-5: bf 31,0f
- lbz 6,4(4)
- stb 6,4(3)
- b 0f
-
-2: bf 30,1f
- lhz 6,0(4)
- sth 6,0(3)
- bf 31,0f
- lbz 6,2(4)
- stb 6,2(3)
- b 0f
-
-1: bf 31,0f
- lbz 6,0(4)
- stb 6,0(3)
-0:
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
-
-/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
- bytes. Each case is handled without loops, using binary (1,2,4,8)
- tests.
-
- In the short (0-8 byte) case no attempt is made to force alignment
- of either source or destination. The hardware will handle the
- unaligned load/stores with small delays for crossing 32- 128-byte,
- and 4096-byte boundaries. Since these short moves are unlikely to be
- unaligned or cross these boundaries, the overhead to force
- alignment is not justified.
-
- The longer (9-31 byte) move is more likely to cross 32- or 128-byte
- boundaries. Since only loads are sensitive to the 32-/128-byte
- boundaries it is more important to align the source then the
- destination. If the source is not already word aligned, we first
- move 1-3 bytes as needed. Since we are only word aligned we don't
- use double word load/stores to insure that all loads are aligned.
- While the destination and stores may still be unaligned, this
- is only an issue for page (4096 byte boundary) crossing, which
- should be rare for these short moves. The hardware handles this
- case automatically with a small (~20 cycle) delay. */
- .align 4
-.L2:
- mtcrf 0x01,5
- neg 8,4
- clrrdi 11,4,2
- andi. 0,8,3
- ble cr6,.LE8 /* Handle moves of 0-8 bytes. */
-/* At least 9 bytes left. Get the source word aligned. */
- cmpldi cr1,5,16
- mr 10,5
- mr 12,4
- cmpldi cr6,0,2
- beq L(dus_tail) /* If the source is already word aligned skip this. */
-/* Copy 1-3 bytes to get source address word aligned. */
- lwz 6,0(11)
- subf 10,0,5
- add 12,4,0
- blt cr6,5f
- srdi 7,6,16
- bgt cr6,3f
-#ifdef __LITTLE_ENDIAN__
- sth 7,0(3)
-#else
- sth 6,0(3)
-#endif
- b 7f
- .align 4
-3:
-#ifdef __LITTLE_ENDIAN__
- rotlwi 6,6,24
- stb 6,0(3)
- sth 7,1(3)
-#else
- stb 7,0(3)
- sth 6,1(3)
-#endif
- b 7f
- .align 4
-5:
-#ifdef __LITTLE_ENDIAN__
- rotlwi 6,6,8
-#endif
- stb 6,0(3)
-7:
- cmpldi cr1,10,16
- add 3,3,0
- mtcrf 0x01,10
- .align 4
-L(dus_tail):
-/* At least 6 bytes left and the source is word aligned. This allows
- some speculative loads up front. */
-/* We need to special case the fall-through because the biggest delays
- are due to address computation not being ready in time for the
- AGEN. */
- lwz 6,0(12)
- lwz 7,4(12)
- blt cr1,L(dus_tail8)
- cmpldi cr0,10,24
-L(dus_tail16): /* Move 16 bytes. */
- stw 6,0(3)
- stw 7,4(3)
- lwz 6,8(12)
- lwz 7,12(12)
- stw 6,8(3)
- stw 7,12(3)
-/* Move 8 bytes more. */
- bf 28,L(dus_tail16p8)
- cmpldi cr1,10,28
- lwz 6,16(12)
- lwz 7,20(12)
- stw 6,16(3)
- stw 7,20(3)
-/* Move 4 bytes more. */
- bf 29,L(dus_tail16p4)
- lwz 6,24(12)
- stw 6,24(3)
- addi 12,12,28
- addi 3,3,28
- bgt cr1,L(dus_tail2)
- /* exactly 28 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail16p8): /* less than 8 bytes left. */
- beq cr1,L(dus_tailX) /* exactly 16 bytes, early exit. */
- cmpldi cr1,10,20
- bf 29,L(dus_tail16p2)
-/* Move 4 bytes more. */
- lwz 6,16(12)
- stw 6,16(3)
- addi 12,12,20
- addi 3,3,20
- bgt cr1,L(dus_tail2)
- /* exactly 20 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail16p4): /* less than 4 bytes left. */
- addi 12,12,24
- addi 3,3,24
- bgt cr0,L(dus_tail2)
- /* exactly 24 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail16p2): /* 16 bytes moved, less than 4 bytes left. */
- addi 12,12,16
- addi 3,3,16
- b L(dus_tail2)
-
- .align 4
-L(dus_tail8): /* Move 8 bytes. */
-/* r6, r7 already loaded speculatively. */
- cmpldi cr1,10,8
- cmpldi cr0,10,12
- bf 28,L(dus_tail4)
- .align 2
- stw 6,0(3)
- stw 7,4(3)
-/* Move 4 bytes more. */
- bf 29,L(dus_tail8p4)
- lwz 6,8(12)
- stw 6,8(3)
- addi 12,12,12
- addi 3,3,12
- bgt cr0,L(dus_tail2)
- /* exactly 12 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail8p4): /* less than 4 bytes left. */
- addi 12,12,8
- addi 3,3,8
- bgt cr1,L(dus_tail2)
- /* exactly 8 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
-
- .align 4
-L(dus_tail4): /* Move 4 bytes. */
-/* r6 already loaded speculatively. If we are here we know there is
- more than 4 bytes left. So there is no need to test. */
- addi 12,12,4
- stw 6,0(3)
- addi 3,3,4
-L(dus_tail2): /* Move 2-3 bytes. */
- bf 30,L(dus_tail1)
- lhz 6,0(12)
- sth 6,0(3)
- bf 31,L(dus_tailX)
- lbz 7,2(12)
- stb 7,2(3)
- ld 3,-16(1)
- blr
-L(dus_tail1): /* Move 1 byte. */
- bf 31,L(dus_tailX)
- lbz 6,0(12)
- stb 6,0(3)
-L(dus_tailX):
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
-
-/* Special case to copy 0-8 bytes. */
- .align 4
-.LE8:
- mr 12,4
- bne cr6,L(dus_4)
-/* Exactly 8 bytes. We may cross a 32-/128-byte boundary and take a ~20
- cycle delay. This case should be rare and any attempt to avoid this
- would take most of 20 cycles any way. */
- ld 6,0(4)
- std 6,0(3)
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_4):
- bf 29,L(dus_tail2)
- lwz 6,0(4)
- stw 6,0(3)
- bf 30,L(dus_5)
- lhz 7,4(4)
- sth 7,4(3)
- bf 31,L(dus_0)
- lbz 8,6(4)
- stb 8,6(3)
- ld 3,-16(1)
- blr
- .align 4
-L(dus_5):
- bf 31,L(dus_0)
- lbz 6,4(4)
- stb 6,4(3)
-L(dus_0):
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
-
- .align 4
-.L6:
- cfi_offset(31,-8)
- mr 12,4
- mr 31,5
- /* Copy doublewords where the destination is aligned but the source is
- not. Use aligned doubleword loads from the source, shifted to realign
- the data, to allow aligned destination stores. */
- addi 11,9,-1 /* loop DW count is one less than total */
- subf 5,10,12 /* Move source addr to previous full double word. */
- cmpldi cr5, 10, 2
- cmpldi cr0, 10, 4
- mr 4,3
- srdi 8,11,2 /* calculate the 32 byte loop count */
- ld 6,0(5) /* pre load 1st full doubleword. */
- mtcrf 0x01,11
- cmpldi cr6,9,4
- mtctr 8
- ld 7,8(5) /* pre load 2nd full doubleword. */
- bge cr0, L(du4_do)
- blt cr5, L(du1_do)
- beq cr5, L(du2_do)
- b L(du3_do)
-
- .align 4
-L(du1_do):
- bf 30,L(du1_1dw)
-
- /* there are at least two DWs to copy */
- /* FIXME: can combine last shift and "or" into "rldimi" */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 8
- sldi 8,6, 64-8
-#else
- sldi 0,7, 8
- srdi 8,6, 64-8
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du1_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du1_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du1_fini) /* if total DWs = 4, then bypass loop */
- b L(du1_loop)
- .align 4
-L(du1_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du1_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du1_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 8
- sldi 8,6, 64-8
-#else
- sldi 0,7, 8
- srdi 8,6, 64-8
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 8
- sldi 8,6, 64-8
-#else
- sldi 0,7, 8
- srdi 8,6, 64-8
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du1_loop)
- .align 4
-L(du1_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du2_do):
- bf 30,L(du2_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 16
- sldi 8,6, 64-16
-#else
- sldi 0,7, 16
- srdi 8,6, 64-16
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du2_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du2_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du2_fini) /* if total DWs = 4, then bypass loop */
- b L(du2_loop)
- .align 4
-L(du2_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du2_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du2_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 16
- sldi 8,6, 64-16
-#else
- sldi 0,7, 16
- srdi 8,6, 64-16
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 16
- sldi 8,6, 64-16
-#else
- sldi 0,7, 16
- srdi 8,6, 64-16
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du2_loop)
- .align 4
-L(du2_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du3_do):
- bf 30,L(du3_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 24
- sldi 8,6, 64-24
-#else
- sldi 0,7, 24
- srdi 8,6, 64-24
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du3_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du3_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du3_fini) /* if total DWs = 4, then bypass loop */
- b L(du3_loop)
- .align 4
-L(du3_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du3_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du3_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 24
- sldi 8,6, 64-24
-#else
- sldi 0,7, 24
- srdi 8,6, 64-24
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 24
- sldi 8,6, 64-24
-#else
- sldi 0,7, 24
- srdi 8,6, 64-24
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du3_loop)
- .align 4
-L(du3_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du4_do):
- cmpldi cr5, 10, 6
- beq cr0, L(du4_dox)
- blt cr5, L(du5_do)
- beq cr5, L(du6_do)
- b L(du7_do)
-L(du4_dox):
- bf 30,L(du4_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 32
- sldi 8,6, 64-32
-#else
- sldi 0,7, 32
- srdi 8,6, 64-32
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du4_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du4_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du4_fini) /* if total DWs = 4, then bypass loop */
- b L(du4_loop)
- .align 4
-L(du4_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du4_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du4_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 32
- sldi 8,6, 64-32
-#else
- sldi 0,7, 32
- srdi 8,6, 64-32
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 32
- sldi 8,6, 64-32
-#else
- sldi 0,7, 32
- srdi 8,6, 64-32
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du4_loop)
- .align 4
-L(du4_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du5_do):
- bf 30,L(du5_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 40
- sldi 8,6, 64-40
-#else
- sldi 0,7, 40
- srdi 8,6, 64-40
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du5_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du5_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du5_fini) /* if total DWs = 4, then bypass loop */
- b L(du5_loop)
- .align 4
-L(du5_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du5_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du5_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 40
- sldi 8,6, 64-40
-#else
- sldi 0,7, 40
- srdi 8,6, 64-40
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 40
- sldi 8,6, 64-40
-#else
- sldi 0,7, 40
- srdi 8,6, 64-40
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du5_loop)
- .align 4
-L(du5_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du6_do):
- bf 30,L(du6_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 48
- sldi 8,6, 64-48
-#else
- sldi 0,7, 48
- srdi 8,6, 64-48
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du6_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du6_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du6_fini) /* if total DWs = 4, then bypass loop */
- b L(du6_loop)
- .align 4
-L(du6_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du6_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du6_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 48
- sldi 8,6, 64-48
-#else
- sldi 0,7, 48
- srdi 8,6, 64-48
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 48
- sldi 8,6, 64-48
-#else
- sldi 0,7, 48
- srdi 8,6, 64-48
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du6_loop)
- .align 4
-L(du6_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du7_do):
- bf 30,L(du7_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 56
- sldi 8,6, 64-56
-#else
- sldi 0,7, 56
- srdi 8,6, 64-56
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du7_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du7_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du7_fini) /* if total DWs = 4, then bypass loop */
- b L(du7_loop)
- .align 4
-L(du7_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du7_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du7_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 56
- sldi 8,6, 64-56
-#else
- sldi 0,7, 56
- srdi 8,6, 64-56
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 56
- sldi 8,6, 64-56
-#else
- sldi 0,7, 56
- srdi 8,6, 64-56
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du7_loop)
- .align 4
-L(du7_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du_done):
- rldicr 0,31,0,60
- mtcrf 0x01,31
- beq cr1,0f /* If the tail is 0 bytes we are done! */
-
- add 3,3,0
- add 12,12,0
-/* At this point we have a tail of 0-7 bytes and we know that the
- destination is double word aligned. */
-4: bf 29,2f
- lwz 6,0(12)
- addi 12,12,4
- stw 6,0(3)
- addi 3,3,4
-2: bf 30,1f
- lhz 6,0(12)
- addi 12,12,2
- sth 6,0(3)
- addi 3,3,2
-1: bf 31,0f
- lbz 6,0(12)
- stb 6,0(3)
-0:
- /* Return original dst pointer. */
- ld 31,-8(1)
- ld 3,-16(1)
- blr
-END_GEN_TB (MEMCPY,TB_TOCLESS)
-libc_hidden_builtin_def (memcpy)
diff --git a/sysdeps/powerpc/powerpc64/power6/memset.S b/sysdeps/powerpc/powerpc64/power6/memset.S
deleted file mode 100644
index aee1c8eabb..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/memset.S
+++ /dev/null
@@ -1,395 +0,0 @@
-/* Optimized 64-bit memset implementation for POWER6.
- Copyright (C) 1997-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-
-/* __ptr_t [r3] memset (__ptr_t s [r3], int c [r4], size_t n [r5]));
- Returns 's'.
-
- The memset is done in three sizes: byte (8 bits), word (32 bits),
- cache line (256 bits). There is a special case for setting cache lines
- to 0, to take advantage of the dcbz instruction. */
-
-#ifndef MEMSET
-# define MEMSET memset
-#endif
- .machine power6
-EALIGN (MEMSET, 7, 0)
- CALL_MCOUNT 3
-
-#define rTMP r0
-#define rRTN r3 /* Initial value of 1st argument. */
-#define rMEMP0 r3 /* Original value of 1st arg. */
-#define rCHR r4 /* Char to set in each byte. */
-#define rLEN r5 /* Length of region to set. */
-#define rMEMP r6 /* Address at which we are storing. */
-#define rALIGN r7 /* Number of bytes we are setting now (when aligning). */
-#define rMEMP2 r8
-#define rMEMP3 r9 /* Alt mem pointer. */
-L(_memset):
-/* Take care of case for size <= 4. */
- cmpldi cr1, rLEN, 8
- andi. rALIGN, rMEMP0, 7
- mr rMEMP, rMEMP0
- ble cr1, L(small)
-
-/* Align to doubleword boundary. */
- cmpldi cr5, rLEN, 31
- insrdi rCHR, rCHR, 8, 48 /* Replicate byte to halfword. */
- beq+ L(aligned2)
- mtcrf 0x01, rMEMP0
- subfic rALIGN, rALIGN, 8
- cror 28,30,31 /* Detect odd word aligned. */
- add rMEMP, rMEMP, rALIGN
- sub rLEN, rLEN, rALIGN
- insrdi rCHR, rCHR, 16, 32 /* Replicate halfword to word. */
- bt 29, L(g4)
-/* Process the even word of doubleword. */
- bf+ 31, L(g2)
- stb rCHR, 0(rMEMP0)
- bt 30, L(g4x)
-L(g2):
- sth rCHR, -6(rMEMP)
-L(g4x):
- stw rCHR, -4(rMEMP)
- b L(aligned)
-/* Process the odd word of doubleword. */
-L(g4):
- bf 28, L(g4x) /* If false, word aligned on odd word. */
- bf+ 31, L(g0)
- stb rCHR, 0(rMEMP0)
- bt 30, L(aligned)
-L(g0):
- sth rCHR, -2(rMEMP)
-
-/* Handle the case of size < 31. */
-L(aligned2):
- insrdi rCHR, rCHR, 16, 32 /* Replicate halfword to word. */
-L(aligned):
- mtcrf 0x01, rLEN
- ble cr5, L(medium)
-/* Align to 32-byte boundary. */
- andi. rALIGN, rMEMP, 0x18
- subfic rALIGN, rALIGN, 0x20
- insrdi rCHR, rCHR, 32, 0 /* Replicate word to double word. */
- beq L(caligned)
- mtcrf 0x01, rALIGN
- add rMEMP, rMEMP, rALIGN
- sub rLEN, rLEN, rALIGN
- cmplwi cr1, rALIGN, 0x10
- mr rMEMP2, rMEMP
- bf 28, L(a1)
- stdu rCHR, -8(rMEMP2)
-L(a1): blt cr1, L(a2)
- std rCHR, -8(rMEMP2)
- stdu rCHR, -16(rMEMP2)
-L(a2):
-
-/* Now aligned to a 32 byte boundary. */
- .align 4
-L(caligned):
- cmpldi cr1, rCHR, 0
- clrrdi. rALIGN, rLEN, 5
- mtcrf 0x01, rLEN
- beq cr1, L(zloopstart) /* Special case for clearing memory using dcbz. */
- beq L(medium) /* We may not actually get to do a full line. */
- .align 4
-/* Storing a non-zero "c" value. We are aligned at a sector (32-byte)
- boundary may not be at cache line (128-byte) boundary. */
-L(nzloopstart):
-/* memset in 32-byte chunks until we get to a cache line boundary.
- If rLEN is less than the distance to the next cache-line boundary use
- cacheAligned1 code to finish the tail. */
- cmpldi cr1,rLEN,128
-
- andi. rTMP,rMEMP,127
- blt cr1,L(cacheAligned1)
- addi rMEMP3,rMEMP,32
- beq L(nzCacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- addi rMEMP,rMEMP,32
- andi. rTMP,rMEMP3,127
- std rCHR,-8(rMEMP3)
-
- beq L(nzCacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP3)
- addi rMEMP,rMEMP,32
- std rCHR,8(rMEMP3)
- andi. rTMP,rMEMP,127
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
-
- beq L(nzCacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,32(rMEMP3)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,128
- std rCHR,40(rMEMP3)
- cmpldi cr6,rLEN,256
- li rMEMP2,128
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
- blt cr1,L(cacheAligned1)
- b L(nzCacheAligned128)
-
-/* Now we are aligned to the cache line and can use dcbtst. */
- .align 4
-L(nzCacheAligned):
- cmpldi cr1,rLEN,128
- blt cr1,L(cacheAligned1)
- b L(nzCacheAligned128)
- .align 5
-L(nzCacheAligned128):
- cmpldi cr1,rLEN,256
- addi rMEMP3,rMEMP,64
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- std rCHR,24(rMEMP)
- std rCHR,32(rMEMP)
- std rCHR,40(rMEMP)
- std rCHR,48(rMEMP)
- std rCHR,56(rMEMP)
- addi rMEMP,rMEMP3,64
- addi rLEN,rLEN,-128
- std rCHR,0(rMEMP3)
- std rCHR,8(rMEMP3)
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
- std rCHR,32(rMEMP3)
- std rCHR,40(rMEMP3)
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
- bge cr1,L(nzCacheAligned128)
- dcbtst 0,rMEMP
- b L(cacheAligned1)
- .align 5
-/* Storing a zero "c" value. We are aligned at a sector (32-byte)
- boundary but may not be at cache line (128-byte) boundary. If the
- remaining length spans a full cache line we can use the Data cache
- block zero instruction. */
-L(zloopstart):
-/* memset in 32-byte chunks until we get to a cache line boundary.
- If rLEN is less than the distance to the next cache-line boundary use
- cacheAligned1 code to finish the tail. */
- cmpldi cr1,rLEN,128
- beq L(medium)
-L(getCacheAligned):
- andi. rTMP,rMEMP,127
- nop
- blt cr1,L(cacheAligned1)
- addi rMEMP3,rMEMP,32
- beq L(cacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- addi rMEMP,rMEMP,32
- andi. rTMP,rMEMP3,127
- std rCHR,-8(rMEMP3)
-L(getCacheAligned2):
- beq L(cacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP3)
- std rCHR,8(rMEMP3)
- addi rMEMP,rMEMP,32
- andi. rTMP,rMEMP,127
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
-L(getCacheAligned3):
- beq L(cacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,32(rMEMP3)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,128
- std rCHR,40(rMEMP3)
- cmpldi cr6,rLEN,256
- li rMEMP2,128
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
- blt cr1,L(cacheAligned1)
- blt cr6,L(cacheAligned128)
- b L(cacheAlignedx)
-
-/* Now we are aligned to the cache line and can use dcbz. */
- .align 5
-L(cacheAligned):
- cmpldi cr1,rLEN,128
- cmpldi cr6,rLEN,256
- blt cr1,L(cacheAligned1)
- li rMEMP2,128
-L(cacheAlignedx):
- cmpldi cr5,rLEN,640
- blt cr6,L(cacheAligned128)
- bgt cr5,L(cacheAligned512)
- cmpldi cr6,rLEN,512
- dcbz 0,rMEMP
- cmpldi cr1,rLEN,384
- dcbz rMEMP2,rMEMP
- addi rMEMP,rMEMP,256
- addi rLEN,rLEN,-256
- blt cr1,L(cacheAligned1)
- blt cr6,L(cacheAligned128)
- b L(cacheAligned256)
- .align 5
-/* A simple loop for the longer (>640 bytes) lengths. This form limits
- the branch miss-predicted to exactly 1 at loop exit.*/
-L(cacheAligned512):
- cmpldi cr1,rLEN,128
- blt cr1,L(cacheAligned1)
- dcbz 0,rMEMP
- addi rLEN,rLEN,-128
- addi rMEMP,rMEMP,128
- b L(cacheAligned512)
- .align 5
-L(cacheAligned256):
-
- cmpldi cr6,rLEN,512
-
- dcbz 0,rMEMP
- cmpldi cr1,rLEN,384
- dcbz rMEMP2,rMEMP
- addi rMEMP,rMEMP,256
- addi rLEN,rLEN,-256
-
- bge cr6,L(cacheAligned256)
-
- blt cr1,L(cacheAligned1)
- .align 4
-L(cacheAligned128):
- dcbz 0,rMEMP
- addi rMEMP,rMEMP,128
- addi rLEN,rLEN,-128
- nop
-L(cacheAligned1):
- cmpldi cr1,rLEN,32
- blt cr1,L(handletail32)
- addi rMEMP3,rMEMP,32
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,32
- std rCHR,-8(rMEMP3)
-L(cacheAligned2):
- blt cr1,L(handletail32)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP3)
- std rCHR,8(rMEMP3)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,32
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
- nop
-L(cacheAligned3):
- blt cr1,L(handletail32)
- addi rMEMP,rMEMP,32
- addi rLEN,rLEN,-32
- std rCHR,32(rMEMP3)
- std rCHR,40(rMEMP3)
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
-
-/* We are here because the length or remainder (rLEN) is less than the
- cache line/sector size and does not justify aggressive loop unrolling.
- So set up the preconditions for L(medium) and go there. */
- .align 3
-L(handletail32):
- cmpldi cr1,rLEN,0
- beqlr cr1
- b L(medium)
-
- .align 5
-L(small):
-/* Memset of 8 bytes or less. */
- cmpldi cr6, rLEN, 4
- cmpldi cr5, rLEN, 1
- ble cr6,L(le4)
- subi rLEN, rLEN, 4
- stb rCHR,0(rMEMP)
- stb rCHR,1(rMEMP)
- stb rCHR,2(rMEMP)
- stb rCHR,3(rMEMP)
- addi rMEMP,rMEMP, 4
- cmpldi cr5, rLEN, 1
-L(le4):
- cmpldi cr1, rLEN, 3
- bltlr cr5
- stb rCHR, 0(rMEMP)
- beqlr cr5
- stb rCHR, 1(rMEMP)
- bltlr cr1
- stb rCHR, 2(rMEMP)
- beqlr cr1
- stb rCHR, 3(rMEMP)
- blr
-
-/* Memset of 0-31 bytes. */
- .align 5
-L(medium):
- insrdi rCHR, rCHR, 32, 0 /* Replicate word to double word. */
- cmpldi cr1, rLEN, 16
-L(medium_tail2):
- add rMEMP, rMEMP, rLEN
-L(medium_tail):
- bt- 31, L(medium_31t)
- bt- 30, L(medium_30t)
-L(medium_30f):
- bt 29, L(medium_29t)
-L(medium_29f):
- bge cr1, L(medium_27t)
- bflr 28
- std rCHR, -8(rMEMP)
- blr
-
-L(medium_31t):
- stbu rCHR, -1(rMEMP)
- bf- 30, L(medium_30f)
-L(medium_30t):
- sthu rCHR, -2(rMEMP)
- bf- 29, L(medium_29f)
-L(medium_29t):
- stwu rCHR, -4(rMEMP)
- blt cr1, L(medium_27f)
-L(medium_27t):
- std rCHR, -8(rMEMP)
- stdu rCHR, -16(rMEMP)
-L(medium_27f):
- bflr 28
-L(medium_28t):
- std rCHR, -8(rMEMP)
- blr
-END_GEN_TB (MEMSET,TB_TOCLESS)
-libc_hidden_builtin_def (memset)
-
-/* Copied from bzero.S to prevent the linker from inserting a stub
- between bzero and memset. */
-ENTRY (__bzero)
- CALL_MCOUNT 3
- mr r5,r4
- li r4,0
- b L(_memset)
-END (__bzero)
-#ifndef __bzero
-weak_alias (__bzero, bzero)
-#endif
diff --git a/sysdeps/powerpc/powerpc64/power6/multiarch/Implies b/sysdeps/powerpc/powerpc64/power6/multiarch/Implies
deleted file mode 100644
index 2ebe304fa6..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/multiarch/Implies
+++ /dev/null
@@ -1 +0,0 @@
-powerpc/powerpc64/power5+/multiarch
diff --git a/sysdeps/powerpc/powerpc64/power6/wcschr.c b/sysdeps/powerpc/powerpc64/power6/wcschr.c
deleted file mode 100644
index ae04a130cc..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/wcschr.c
+++ /dev/null
@@ -1 +0,0 @@
-#include <sysdeps/powerpc/power6/wcschr.c>
diff --git a/sysdeps/powerpc/powerpc64/power6/wcscpy.c b/sysdeps/powerpc/powerpc64/power6/wcscpy.c
deleted file mode 100644
index 722c8f995b..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/wcscpy.c
+++ /dev/null
@@ -1 +0,0 @@
-#include <sysdeps/powerpc/power6/wcscpy.c>
diff --git a/sysdeps/powerpc/powerpc64/power6/wcsrchr.c b/sysdeps/powerpc/powerpc64/power6/wcsrchr.c
deleted file mode 100644
index b86472d7bd..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/wcsrchr.c
+++ /dev/null
@@ -1 +0,0 @@
-#include <sysdeps/powerpc/power6/wcsrchr.c>