aboutsummaryrefslogtreecommitdiff
path: root/sysdeps/powerpc/powerpc64/power6
diff options
context:
space:
mode:
Diffstat (limited to 'sysdeps/powerpc/powerpc64/power6')
-rw-r--r--sysdeps/powerpc/powerpc64/power6/Implies2
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/Implies1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S58
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S59
-rw-r--r--sysdeps/powerpc/powerpc64/power6/memcpy.S1499
-rw-r--r--sysdeps/powerpc/powerpc64/power6/memset.S395
-rw-r--r--sysdeps/powerpc/powerpc64/power6/multiarch/Implies1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/wcschr.c1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/wcscpy.c1
-rw-r--r--sysdeps/powerpc/powerpc64/power6/wcsrchr.c1
12 files changed, 0 insertions, 2020 deletions
diff --git a/sysdeps/powerpc/powerpc64/power6/Implies b/sysdeps/powerpc/powerpc64/power6/Implies
deleted file mode 100644
index 4c782d4122..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/Implies
+++ /dev/null
@@ -1,2 +0,0 @@
-powerpc/powerpc64/power5+/fpu
-powerpc/powerpc64/power5+
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/Implies b/sysdeps/powerpc/powerpc64/power6/fpu/Implies
deleted file mode 100644
index f09854edb6..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/Implies
+++ /dev/null
@@ -1 +0,0 @@
-powerpc/powerpc64/power5+/fpu
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies b/sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies
deleted file mode 100644
index fca8a4ef0f..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/multiarch/Implies
+++ /dev/null
@@ -1 +0,0 @@
-powerpc/powerpc64/power5+/fpu/multiarch
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S b/sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S
deleted file mode 100644
index ec36d1be5b..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysign.S
+++ /dev/null
@@ -1,58 +0,0 @@
-/* copysign(). PowerPC64/POWER6 version.
- Copyright (C) 2010-2017 Free Software Foundation, Inc.
- Contributed by Luis Machado <luisgpm@br.ibm.com>.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-#include <math_ldbl_opt.h>
-
-/* double [f1] copysign (double [f1] x, double [f2] y);
- copysign(x,y) returns a value with the magnitude of x and
- with the sign bit of y. */
-
- .section ".text"
- .type __copysign, @function
- .machine power6
-EALIGN (__copysign, 4, 0)
- CALL_MCOUNT 0
- fcpsgn fp1,fp2,fp1
- blr
-END (__copysign)
-
-hidden_def (__copysign)
-weak_alias (__copysign, copysign)
-
-/* It turns out that the 'double' version will also always work for
- single-precision. */
-strong_alias (__copysign, __copysignf)
-hidden_def (__copysignf)
-weak_alias (__copysignf, copysignf)
-
-#ifdef NO_LONG_DOUBLE
-strong_alias (__copysign, __copysignl)
-weak_alias (__copysign, copysignl)
-#endif
-
-#if IS_IN (libm)
-# if LONG_DOUBLE_COMPAT (libm, GLIBC_2_0)
-compat_symbol (libm, copysign, copysignl, GLIBC_2_0)
-# endif
-#else
-# if LONG_DOUBLE_COMPAT(libc, GLIBC_2_0)
-compat_symbol (libc, copysign, copysignl, GLIBC_2_0);
-# endif
-#endif
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S b/sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S
deleted file mode 100644
index d4aa702d07..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/s_copysignf.S
+++ /dev/null
@@ -1 +0,0 @@
-/* This function uses the same code as s_copysign.S. */
diff --git a/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S b/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
deleted file mode 100644
index 85187b45f3..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
+++ /dev/null
@@ -1,59 +0,0 @@
-/* isnan(). PowerPC64 version.
- Copyright (C) 2008-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-#include <math_ldbl_opt.h>
-
-/* int __isnan(x) */
- .machine power6
-EALIGN (__isnan, 4, 0)
- CALL_MCOUNT 0
- stfd fp1,-8(r1) /* copy FPR to GPR */
- ori r1,r1,0
- ld r4,-8(r1)
- lis r0,0x7ff0
- sldi r0,r0,32 /* const long r0 0x7ff00000 00000000 */
- clrldi r4,r4,1 /* x = fabs(x) */
- cmpd cr7,r4,r0 /* if (fabs(x) <= inf) */
- li r3,0 /* then return 0 */
- blelr+ cr7
- li r3,1 /* else return 1 */
- blr
- END (__isnan)
-
-hidden_def (__isnan)
-weak_alias (__isnan, isnan)
-
-/* It turns out that the 'double' version will also always work for
- single-precision. */
-strong_alias (__isnan, __isnanf)
-hidden_def (__isnanf)
-weak_alias (__isnanf, isnanf)
-
-#ifdef NO_LONG_DOUBLE
-strong_alias (__isnan, __isnanl)
-weak_alias (__isnan, isnanl)
-#endif
-
-#if !IS_IN (libm)
-# if LONG_DOUBLE_COMPAT(libc, GLIBC_2_0)
-compat_symbol (libc, __isnan, __isnanl, GLIBC_2_0);
-compat_symbol (libc, isnan, isnanl, GLIBC_2_0);
-# endif
-#endif
-
diff --git a/sysdeps/powerpc/powerpc64/power6/memcpy.S b/sysdeps/powerpc/powerpc64/power6/memcpy.S
deleted file mode 100644
index 1f7294b8ed..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/memcpy.S
+++ /dev/null
@@ -1,1499 +0,0 @@
-/* Optimized memcpy implementation for PowerPC64.
- Copyright (C) 2003-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-
-/* __ptr_t [r3] memcpy (__ptr_t dst [r3], __ptr_t src [r4], size_t len [r5]);
- Returns 'dst'.
-
- Memcpy handles short copies (< 32-bytes) using a binary move blocks
- (no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
- with the appropriate combination of byte and halfword load/stores.
- There is minimal effort to optimize the alignment of short moves.
- The 64-bit implementations of POWER3 and POWER4 do a reasonable job
- of handling unaligned load/stores that do not cross 32-byte boundaries.
-
- Longer moves (>= 32-bytes) justify the effort to get at least the
- destination doubleword (8-byte) aligned. Further optimization is
- possible when both source and destination are doubleword aligned.
- Each case has a optimized unrolled loop.
-
- For POWER6 unaligned loads will take a 20+ cycle hiccup for any
- L1 cache miss that crosses a 32- or 128-byte boundary. Store
- is more forgiving and does not take a hiccup until page or
- segment boundaries. So we require doubleword alignment for
- the source but may take a risk and only require word alignment
- for the destination. */
-
-#ifndef MEMCPY
-# define MEMCPY memcpy
-#endif
- .machine "power6"
-EALIGN (MEMCPY, 7, 0)
- CALL_MCOUNT 3
-
- cmpldi cr1,5,31
- neg 0,3
- std 3,-16(1)
- std 31,-8(1)
- andi. 11,3,7 /* check alignment of dst. */
- clrldi 0,0,61 /* Number of bytes until the 1st doubleword of dst. */
- clrldi 10,4,61 /* check alignment of src. */
- cmpldi cr6,5,8
- ble- cr1,.L2 /* If move < 32 bytes use short move code. */
- mtcrf 0x01,0
- cmpld cr6,10,11
- srdi 9,5,3 /* Number of full double words remaining. */
- beq .L0
-
- subf 5,0,5
- /* Move 0-7 bytes as needed to get the destination doubleword aligned.
- Duplicate some code to maximize fall-through and minimize agen delays. */
-1: bf 31,2f
- lbz 6,0(4)
- stb 6,0(3)
- bf 30,5f
- lhz 6,1(4)
- sth 6,1(3)
- bf 29,0f
- lwz 6,3(4)
- stw 6,3(3)
- b 0f
-5:
- bf 29,0f
- lwz 6,1(4)
- stw 6,1(3)
- b 0f
-
-2: bf 30,4f
- lhz 6,0(4)
- sth 6,0(3)
- bf 29,0f
- lwz 6,2(4)
- stw 6,2(3)
- b 0f
-
-4: bf 29,0f
- lwz 6,0(4)
- stw 6,0(3)
-0:
-/* Add the number of bytes until the 1st doubleword of dst to src and dst. */
- add 4,4,0
- add 3,3,0
-
- clrldi 10,4,61 /* check alignment of src again. */
- srdi 9,5,3 /* Number of full double words remaining. */
-
- /* Copy doublewords from source to destination, assuming the
- destination is aligned on a doubleword boundary.
-
- At this point we know there are at least 25 bytes left (32-7) to copy.
- The next step is to determine if the source is also doubleword aligned.
- If not branch to the unaligned move code at .L6. which uses
- a load, shift, store strategy.
-
- Otherwise source and destination are doubleword aligned, and we can
- the optimized doubleword copy loop. */
- .align 4
-.L0:
- clrldi 11,5,61
- andi. 0,5,0x78
- srdi 12,5,7 /* Number of 128-byte blocks to move. */
- cmpldi cr1,11,0 /* If the tail is 0 bytes */
- bne- cr6,.L6 /* If source is not DW aligned. */
-
- /* Move doublewords where destination and source are DW aligned.
- Use a unrolled loop to copy 16 doublewords (128-bytes) per iteration.
- If the copy is not an exact multiple of 128 bytes, 1-15
- doublewords are copied as needed to set up the main loop. After
- the main loop exits there may be a tail of 1-7 bytes. These byte
- are copied a word/halfword/byte at a time as needed to preserve
- alignment.
-
- For POWER6 the L1 is store-through and the L2 is store-in. The
- L2 is clocked at half CPU clock so we can store 16 bytes every
- other cycle. POWER6 also has a load/store bypass so we can do
- load, load, store, store every 2 cycles.
-
- The following code is sensitive to cache line alignment. Do not
- make any change with out first making sure they don't result in
- splitting ld/std pairs across a cache line. */
-
- mtcrf 0x02,5
- mtcrf 0x01,5
- cmpldi cr5,12,1
- beq L(das_loop)
-
- bf 25,4f
- .align 3
- ld 6,0(4)
- ld 7,8(4)
- mr 11,4
- mr 10,3
- std 6,0(3)
- std 7,8(3)
- ld 6,16(4)
- ld 7,24(4)
- std 6,16(3)
- std 7,24(3)
- ld 6,0+32(4)
- ld 7,8+32(4)
- addi 4,4,64
- addi 3,3,64
- std 6,0+32(10)
- std 7,8+32(10)
- ld 6,16+32(11)
- ld 7,24+32(11)
- std 6,16+32(10)
- std 7,24+32(10)
-4:
- mr 10,3
- bf 26,2f
- ld 6,0(4)
- ld 7,8(4)
- mr 11,4
- nop
- std 6,0(3)
- std 7,8(3)
- ld 6,16(4)
- ld 7,24(4)
- addi 4,4,32
- std 6,16(3)
- std 7,24(3)
- addi 3,3,32
-6:
- nop
- bf 27,5f
- ld 6,0+32(11)
- ld 7,8+32(11)
- addi 4,4,16
- addi 3,3,16
- std 6,0+32(10)
- std 7,8+32(10)
- bf 28,L(das_loop_s)
- ld 0,16+32(11)
- addi 4,4,8
- addi 3,3,8
- std 0,16+32(10)
- blt cr5,L(das_tail)
- b L(das_loop)
- .align 3
-5:
- nop
- bf 28,L(das_loop_s)
- ld 6,32(11)
- addi 4,4,8
- addi 3,3,8
- std 6,32(10)
- blt cr5,L(das_tail)
- b L(das_loop)
- .align 3
-2:
- mr 11,4
- bf 27,1f
- ld 6,0(4)
- ld 7,8(4)
- addi 4,4,16
- addi 3,3,16
- std 6,0(10)
- std 7,8(10)
- bf 28,L(das_loop_s)
- ld 0,16(11)
- addi 4,11,24
- addi 3,10,24
- std 0,16(10)
- blt cr5,L(das_tail)
- b L(das_loop)
- .align 3
-1:
- nop
- bf 28,L(das_loop_s)
- ld 6,0(4)
- addi 4,4,8
- addi 3,3,8
- std 6,0(10)
-L(das_loop_s):
- nop
- blt cr5,L(das_tail)
- .align 4
-L(das_loop):
- ld 6,0(4)
- ld 7,8(4)
- mr 10,3
- mr 11,4
- std 6,0(3)
- std 7,8(3)
- addi 12,12,-1
- nop
- ld 8,16(4)
- ld 0,24(4)
- std 8,16(3)
- std 0,24(3)
-
- ld 6,0+32(4)
- ld 7,8+32(4)
- std 6,0+32(3)
- std 7,8+32(3)
- ld 8,16+32(4)
- ld 0,24+32(4)
- std 8,16+32(3)
- std 0,24+32(3)
-
- ld 6,0+64(11)
- ld 7,8+64(11)
- std 6,0+64(10)
- std 7,8+64(10)
- ld 8,16+64(11)
- ld 0,24+64(11)
- std 8,16+64(10)
- std 0,24+64(10)
-
- ld 6,0+96(11)
- ld 7,8+96(11)
- addi 4,4,128
- addi 3,3,128
- std 6,0+96(10)
- std 7,8+96(10)
- ld 8,16+96(11)
- ld 0,24+96(11)
- std 8,16+96(10)
- std 0,24+96(10)
- ble cr5,L(das_loop_e)
-
- mtctr 12
- .align 4
-L(das_loop2):
- ld 6,0(4)
- ld 7,8(4)
- mr 10,3
- mr 11,4
- std 6,0(3)
- std 7,8(3)
- ld 8,16(4)
- ld 0,24(4)
- std 8,16(3)
- std 0,24(3)
-
- ld 6,0+32(4)
- ld 7,8+32(4)
- std 6,0+32(3)
- std 7,8+32(3)
- ld 8,16+32(4)
- ld 0,24+32(4)
- std 8,16+32(3)
- std 0,24+32(3)
-
- ld 6,0+64(11)
- ld 7,8+64(11)
- std 6,0+64(10)
- std 7,8+64(10)
- ld 8,16+64(11)
- ld 0,24+64(11)
- std 8,16+64(10)
- std 0,24+64(10)
-
- ld 6,0+96(11)
- ld 7,8+96(11)
- addi 4,4,128
- addi 3,3,128
- std 6,0+96(10)
- std 7,8+96(10)
- ld 8,16+96(11)
- ld 0,24+96(11)
- std 8,16+96(10)
- std 0,24+96(10)
- bdnz L(das_loop2)
-L(das_loop_e):
-/* Check of a 1-7 byte tail, return if none. */
- bne cr1,L(das_tail2)
-/* Return original dst pointer. */
- ld 3,-16(1)
- blr
- .align 4
-L(das_tail):
- beq cr1,0f
-
-L(das_tail2):
-/* At this point we have a tail of 0-7 bytes and we know that the
- destination is double word aligned. */
-4: bf 29,2f
- lwz 6,0(4)
- stw 6,0(3)
- bf 30,5f
- lhz 6,4(4)
- sth 6,4(3)
- bf 31,0f
- lbz 6,6(4)
- stb 6,6(3)
- b 0f
-5: bf 31,0f
- lbz 6,4(4)
- stb 6,4(3)
- b 0f
-
-2: bf 30,1f
- lhz 6,0(4)
- sth 6,0(3)
- bf 31,0f
- lbz 6,2(4)
- stb 6,2(3)
- b 0f
-
-1: bf 31,0f
- lbz 6,0(4)
- stb 6,0(3)
-0:
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
-
-/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
- bytes. Each case is handled without loops, using binary (1,2,4,8)
- tests.
-
- In the short (0-8 byte) case no attempt is made to force alignment
- of either source or destination. The hardware will handle the
- unaligned load/stores with small delays for crossing 32- 128-byte,
- and 4096-byte boundaries. Since these short moves are unlikely to be
- unaligned or cross these boundaries, the overhead to force
- alignment is not justified.
-
- The longer (9-31 byte) move is more likely to cross 32- or 128-byte
- boundaries. Since only loads are sensitive to the 32-/128-byte
- boundaries it is more important to align the source then the
- destination. If the source is not already word aligned, we first
- move 1-3 bytes as needed. Since we are only word aligned we don't
- use double word load/stores to insure that all loads are aligned.
- While the destination and stores may still be unaligned, this
- is only an issue for page (4096 byte boundary) crossing, which
- should be rare for these short moves. The hardware handles this
- case automatically with a small (~20 cycle) delay. */
- .align 4
-.L2:
- mtcrf 0x01,5
- neg 8,4
- clrrdi 11,4,2
- andi. 0,8,3
- ble cr6,.LE8 /* Handle moves of 0-8 bytes. */
-/* At least 9 bytes left. Get the source word aligned. */
- cmpldi cr1,5,16
- mr 10,5
- mr 12,4
- cmpldi cr6,0,2
- beq L(dus_tail) /* If the source is already word aligned skip this. */
-/* Copy 1-3 bytes to get source address word aligned. */
- lwz 6,0(11)
- subf 10,0,5
- add 12,4,0
- blt cr6,5f
- srdi 7,6,16
- bgt cr6,3f
-#ifdef __LITTLE_ENDIAN__
- sth 7,0(3)
-#else
- sth 6,0(3)
-#endif
- b 7f
- .align 4
-3:
-#ifdef __LITTLE_ENDIAN__
- rotlwi 6,6,24
- stb 6,0(3)
- sth 7,1(3)
-#else
- stb 7,0(3)
- sth 6,1(3)
-#endif
- b 7f
- .align 4
-5:
-#ifdef __LITTLE_ENDIAN__
- rotlwi 6,6,8
-#endif
- stb 6,0(3)
-7:
- cmpldi cr1,10,16
- add 3,3,0
- mtcrf 0x01,10
- .align 4
-L(dus_tail):
-/* At least 6 bytes left and the source is word aligned. This allows
- some speculative loads up front. */
-/* We need to special case the fall-through because the biggest delays
- are due to address computation not being ready in time for the
- AGEN. */
- lwz 6,0(12)
- lwz 7,4(12)
- blt cr1,L(dus_tail8)
- cmpldi cr0,10,24
-L(dus_tail16): /* Move 16 bytes. */
- stw 6,0(3)
- stw 7,4(3)
- lwz 6,8(12)
- lwz 7,12(12)
- stw 6,8(3)
- stw 7,12(3)
-/* Move 8 bytes more. */
- bf 28,L(dus_tail16p8)
- cmpldi cr1,10,28
- lwz 6,16(12)
- lwz 7,20(12)
- stw 6,16(3)
- stw 7,20(3)
-/* Move 4 bytes more. */
- bf 29,L(dus_tail16p4)
- lwz 6,24(12)
- stw 6,24(3)
- addi 12,12,28
- addi 3,3,28
- bgt cr1,L(dus_tail2)
- /* exactly 28 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail16p8): /* less than 8 bytes left. */
- beq cr1,L(dus_tailX) /* exactly 16 bytes, early exit. */
- cmpldi cr1,10,20
- bf 29,L(dus_tail16p2)
-/* Move 4 bytes more. */
- lwz 6,16(12)
- stw 6,16(3)
- addi 12,12,20
- addi 3,3,20
- bgt cr1,L(dus_tail2)
- /* exactly 20 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail16p4): /* less than 4 bytes left. */
- addi 12,12,24
- addi 3,3,24
- bgt cr0,L(dus_tail2)
- /* exactly 24 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail16p2): /* 16 bytes moved, less than 4 bytes left. */
- addi 12,12,16
- addi 3,3,16
- b L(dus_tail2)
-
- .align 4
-L(dus_tail8): /* Move 8 bytes. */
-/* r6, r7 already loaded speculatively. */
- cmpldi cr1,10,8
- cmpldi cr0,10,12
- bf 28,L(dus_tail4)
- .align 2
- stw 6,0(3)
- stw 7,4(3)
-/* Move 4 bytes more. */
- bf 29,L(dus_tail8p4)
- lwz 6,8(12)
- stw 6,8(3)
- addi 12,12,12
- addi 3,3,12
- bgt cr0,L(dus_tail2)
- /* exactly 12 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_tail8p4): /* less than 4 bytes left. */
- addi 12,12,8
- addi 3,3,8
- bgt cr1,L(dus_tail2)
- /* exactly 8 bytes. Return original dst pointer and exit. */
- ld 3,-16(1)
- blr
-
- .align 4
-L(dus_tail4): /* Move 4 bytes. */
-/* r6 already loaded speculatively. If we are here we know there is
- more than 4 bytes left. So there is no need to test. */
- addi 12,12,4
- stw 6,0(3)
- addi 3,3,4
-L(dus_tail2): /* Move 2-3 bytes. */
- bf 30,L(dus_tail1)
- lhz 6,0(12)
- sth 6,0(3)
- bf 31,L(dus_tailX)
- lbz 7,2(12)
- stb 7,2(3)
- ld 3,-16(1)
- blr
-L(dus_tail1): /* Move 1 byte. */
- bf 31,L(dus_tailX)
- lbz 6,0(12)
- stb 6,0(3)
-L(dus_tailX):
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
-
-/* Special case to copy 0-8 bytes. */
- .align 4
-.LE8:
- mr 12,4
- bne cr6,L(dus_4)
-/* Exactly 8 bytes. We may cross a 32-/128-byte boundary and take a ~20
- cycle delay. This case should be rare and any attempt to avoid this
- would take most of 20 cycles any way. */
- ld 6,0(4)
- std 6,0(3)
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
- .align 4
-L(dus_4):
- bf 29,L(dus_tail2)
- lwz 6,0(4)
- stw 6,0(3)
- bf 30,L(dus_5)
- lhz 7,4(4)
- sth 7,4(3)
- bf 31,L(dus_0)
- lbz 8,6(4)
- stb 8,6(3)
- ld 3,-16(1)
- blr
- .align 4
-L(dus_5):
- bf 31,L(dus_0)
- lbz 6,4(4)
- stb 6,4(3)
-L(dus_0):
- /* Return original dst pointer. */
- ld 3,-16(1)
- blr
-
- .align 4
-.L6:
- cfi_offset(31,-8)
- mr 12,4
- mr 31,5
- /* Copy doublewords where the destination is aligned but the source is
- not. Use aligned doubleword loads from the source, shifted to realign
- the data, to allow aligned destination stores. */
- addi 11,9,-1 /* loop DW count is one less than total */
- subf 5,10,12 /* Move source addr to previous full double word. */
- cmpldi cr5, 10, 2
- cmpldi cr0, 10, 4
- mr 4,3
- srdi 8,11,2 /* calculate the 32 byte loop count */
- ld 6,0(5) /* pre load 1st full doubleword. */
- mtcrf 0x01,11
- cmpldi cr6,9,4
- mtctr 8
- ld 7,8(5) /* pre load 2nd full doubleword. */
- bge cr0, L(du4_do)
- blt cr5, L(du1_do)
- beq cr5, L(du2_do)
- b L(du3_do)
-
- .align 4
-L(du1_do):
- bf 30,L(du1_1dw)
-
- /* there are at least two DWs to copy */
- /* FIXME: can combine last shift and "or" into "rldimi" */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 8
- sldi 8,6, 64-8
-#else
- sldi 0,7, 8
- srdi 8,6, 64-8
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du1_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du1_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du1_fini) /* if total DWs = 4, then bypass loop */
- b L(du1_loop)
- .align 4
-L(du1_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du1_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du1_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 8
- sldi 8,6, 64-8
-#else
- sldi 0,7, 8
- srdi 8,6, 64-8
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 8
- sldi 8,6, 64-8
-#else
- sldi 0,7, 8
- srdi 8,6, 64-8
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du1_loop)
- .align 4
-L(du1_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 8
- sldi 8,7, 64-8
-#else
- sldi 0,6, 8
- srdi 8,7, 64-8
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du2_do):
- bf 30,L(du2_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 16
- sldi 8,6, 64-16
-#else
- sldi 0,7, 16
- srdi 8,6, 64-16
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du2_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du2_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du2_fini) /* if total DWs = 4, then bypass loop */
- b L(du2_loop)
- .align 4
-L(du2_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du2_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du2_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 16
- sldi 8,6, 64-16
-#else
- sldi 0,7, 16
- srdi 8,6, 64-16
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 16
- sldi 8,6, 64-16
-#else
- sldi 0,7, 16
- srdi 8,6, 64-16
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du2_loop)
- .align 4
-L(du2_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 16
- sldi 8,7, 64-16
-#else
- sldi 0,6, 16
- srdi 8,7, 64-16
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du3_do):
- bf 30,L(du3_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 24
- sldi 8,6, 64-24
-#else
- sldi 0,7, 24
- srdi 8,6, 64-24
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du3_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du3_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du3_fini) /* if total DWs = 4, then bypass loop */
- b L(du3_loop)
- .align 4
-L(du3_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du3_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du3_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 24
- sldi 8,6, 64-24
-#else
- sldi 0,7, 24
- srdi 8,6, 64-24
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 24
- sldi 8,6, 64-24
-#else
- sldi 0,7, 24
- srdi 8,6, 64-24
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du3_loop)
- .align 4
-L(du3_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 24
- sldi 8,7, 64-24
-#else
- sldi 0,6, 24
- srdi 8,7, 64-24
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du4_do):
- cmpldi cr5, 10, 6
- beq cr0, L(du4_dox)
- blt cr5, L(du5_do)
- beq cr5, L(du6_do)
- b L(du7_do)
-L(du4_dox):
- bf 30,L(du4_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 32
- sldi 8,6, 64-32
-#else
- sldi 0,7, 32
- srdi 8,6, 64-32
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du4_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du4_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du4_fini) /* if total DWs = 4, then bypass loop */
- b L(du4_loop)
- .align 4
-L(du4_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du4_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du4_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 32
- sldi 8,6, 64-32
-#else
- sldi 0,7, 32
- srdi 8,6, 64-32
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 32
- sldi 8,6, 64-32
-#else
- sldi 0,7, 32
- srdi 8,6, 64-32
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du4_loop)
- .align 4
-L(du4_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 32
- sldi 8,7, 64-32
-#else
- sldi 0,6, 32
- srdi 8,7, 64-32
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du5_do):
- bf 30,L(du5_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 40
- sldi 8,6, 64-40
-#else
- sldi 0,7, 40
- srdi 8,6, 64-40
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du5_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du5_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du5_fini) /* if total DWs = 4, then bypass loop */
- b L(du5_loop)
- .align 4
-L(du5_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du5_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du5_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 40
- sldi 8,6, 64-40
-#else
- sldi 0,7, 40
- srdi 8,6, 64-40
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 40
- sldi 8,6, 64-40
-#else
- sldi 0,7, 40
- srdi 8,6, 64-40
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du5_loop)
- .align 4
-L(du5_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 40
- sldi 8,7, 64-40
-#else
- sldi 0,6, 40
- srdi 8,7, 64-40
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du6_do):
- bf 30,L(du6_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 48
- sldi 8,6, 64-48
-#else
- sldi 0,7, 48
- srdi 8,6, 64-48
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du6_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du6_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du6_fini) /* if total DWs = 4, then bypass loop */
- b L(du6_loop)
- .align 4
-L(du6_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du6_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du6_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 48
- sldi 8,6, 64-48
-#else
- sldi 0,7, 48
- srdi 8,6, 64-48
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 48
- sldi 8,6, 64-48
-#else
- sldi 0,7, 48
- srdi 8,6, 64-48
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du6_loop)
- .align 4
-L(du6_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 48
- sldi 8,7, 64-48
-#else
- sldi 0,6, 48
- srdi 8,7, 64-48
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du7_do):
- bf 30,L(du7_1dw)
-
- /* there are at least two DWs to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 56
- sldi 8,6, 64-56
-#else
- sldi 0,7, 56
- srdi 8,6, 64-56
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,8(4)
- addi 4,4,16
- addi 5,5,32
- blt cr6,L(du7_fini) /* if total DWs = 3, then bypass loop */
- bf 31,L(du7_loop)
- /* there is a third DW to copy */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- std 0,0(4)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- addi 4,4,8
- beq cr6,L(du7_fini) /* if total DWs = 4, then bypass loop */
- b L(du7_loop)
- .align 4
-L(du7_1dw):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- addi 5,5,16
- or 0,0,8
- bf 31,L(du7_loop)
- mr 6,7
- ld 7,0(5)
- addi 5,5,8
- std 0,0(4)
- addi 4,4,8
- .align 4
-/* copy 32 bytes at a time */
-L(du7_loop):
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- ld 6,0(5)
- std 0,0(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 56
- sldi 8,6, 64-56
-#else
- sldi 0,7, 56
- srdi 8,6, 64-56
-#endif
- or 0,0,8
- ld 7,8(5)
- std 0,8(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- ld 6,16(5)
- std 0,16(4)
-#ifdef __LITTLE_ENDIAN__
- srdi 0,7, 56
- sldi 8,6, 64-56
-#else
- sldi 0,7, 56
- srdi 8,6, 64-56
-#endif
- or 0,0,8
- ld 7,24(5)
- std 0,24(4)
- addi 5,5,32
- addi 4,4,32
- bdnz+ L(du7_loop)
- .align 4
-L(du7_fini):
- /* calculate and store the final DW */
-#ifdef __LITTLE_ENDIAN__
- srdi 0,6, 56
- sldi 8,7, 64-56
-#else
- sldi 0,6, 56
- srdi 8,7, 64-56
-#endif
- or 0,0,8
- std 0,0(4)
- b L(du_done)
-
- .align 4
-L(du_done):
- rldicr 0,31,0,60
- mtcrf 0x01,31
- beq cr1,0f /* If the tail is 0 bytes we are done! */
-
- add 3,3,0
- add 12,12,0
-/* At this point we have a tail of 0-7 bytes and we know that the
- destination is double word aligned. */
-4: bf 29,2f
- lwz 6,0(12)
- addi 12,12,4
- stw 6,0(3)
- addi 3,3,4
-2: bf 30,1f
- lhz 6,0(12)
- addi 12,12,2
- sth 6,0(3)
- addi 3,3,2
-1: bf 31,0f
- lbz 6,0(12)
- stb 6,0(3)
-0:
- /* Return original dst pointer. */
- ld 31,-8(1)
- ld 3,-16(1)
- blr
-END_GEN_TB (MEMCPY,TB_TOCLESS)
-libc_hidden_builtin_def (memcpy)
diff --git a/sysdeps/powerpc/powerpc64/power6/memset.S b/sysdeps/powerpc/powerpc64/power6/memset.S
deleted file mode 100644
index aee1c8eabb..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/memset.S
+++ /dev/null
@@ -1,395 +0,0 @@
-/* Optimized 64-bit memset implementation for POWER6.
- Copyright (C) 1997-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-
-/* __ptr_t [r3] memset (__ptr_t s [r3], int c [r4], size_t n [r5]));
- Returns 's'.
-
- The memset is done in three sizes: byte (8 bits), word (32 bits),
- cache line (256 bits). There is a special case for setting cache lines
- to 0, to take advantage of the dcbz instruction. */
-
-#ifndef MEMSET
-# define MEMSET memset
-#endif
- .machine power6
-EALIGN (MEMSET, 7, 0)
- CALL_MCOUNT 3
-
-#define rTMP r0
-#define rRTN r3 /* Initial value of 1st argument. */
-#define rMEMP0 r3 /* Original value of 1st arg. */
-#define rCHR r4 /* Char to set in each byte. */
-#define rLEN r5 /* Length of region to set. */
-#define rMEMP r6 /* Address at which we are storing. */
-#define rALIGN r7 /* Number of bytes we are setting now (when aligning). */
-#define rMEMP2 r8
-#define rMEMP3 r9 /* Alt mem pointer. */
-L(_memset):
-/* Take care of case for size <= 4. */
- cmpldi cr1, rLEN, 8
- andi. rALIGN, rMEMP0, 7
- mr rMEMP, rMEMP0
- ble cr1, L(small)
-
-/* Align to doubleword boundary. */
- cmpldi cr5, rLEN, 31
- insrdi rCHR, rCHR, 8, 48 /* Replicate byte to halfword. */
- beq+ L(aligned2)
- mtcrf 0x01, rMEMP0
- subfic rALIGN, rALIGN, 8
- cror 28,30,31 /* Detect odd word aligned. */
- add rMEMP, rMEMP, rALIGN
- sub rLEN, rLEN, rALIGN
- insrdi rCHR, rCHR, 16, 32 /* Replicate halfword to word. */
- bt 29, L(g4)
-/* Process the even word of doubleword. */
- bf+ 31, L(g2)
- stb rCHR, 0(rMEMP0)
- bt 30, L(g4x)
-L(g2):
- sth rCHR, -6(rMEMP)
-L(g4x):
- stw rCHR, -4(rMEMP)
- b L(aligned)
-/* Process the odd word of doubleword. */
-L(g4):
- bf 28, L(g4x) /* If false, word aligned on odd word. */
- bf+ 31, L(g0)
- stb rCHR, 0(rMEMP0)
- bt 30, L(aligned)
-L(g0):
- sth rCHR, -2(rMEMP)
-
-/* Handle the case of size < 31. */
-L(aligned2):
- insrdi rCHR, rCHR, 16, 32 /* Replicate halfword to word. */
-L(aligned):
- mtcrf 0x01, rLEN
- ble cr5, L(medium)
-/* Align to 32-byte boundary. */
- andi. rALIGN, rMEMP, 0x18
- subfic rALIGN, rALIGN, 0x20
- insrdi rCHR, rCHR, 32, 0 /* Replicate word to double word. */
- beq L(caligned)
- mtcrf 0x01, rALIGN
- add rMEMP, rMEMP, rALIGN
- sub rLEN, rLEN, rALIGN
- cmplwi cr1, rALIGN, 0x10
- mr rMEMP2, rMEMP
- bf 28, L(a1)
- stdu rCHR, -8(rMEMP2)
-L(a1): blt cr1, L(a2)
- std rCHR, -8(rMEMP2)
- stdu rCHR, -16(rMEMP2)
-L(a2):
-
-/* Now aligned to a 32 byte boundary. */
- .align 4
-L(caligned):
- cmpldi cr1, rCHR, 0
- clrrdi. rALIGN, rLEN, 5
- mtcrf 0x01, rLEN
- beq cr1, L(zloopstart) /* Special case for clearing memory using dcbz. */
- beq L(medium) /* We may not actually get to do a full line. */
- .align 4
-/* Storing a non-zero "c" value. We are aligned at a sector (32-byte)
- boundary may not be at cache line (128-byte) boundary. */
-L(nzloopstart):
-/* memset in 32-byte chunks until we get to a cache line boundary.
- If rLEN is less than the distance to the next cache-line boundary use
- cacheAligned1 code to finish the tail. */
- cmpldi cr1,rLEN,128
-
- andi. rTMP,rMEMP,127
- blt cr1,L(cacheAligned1)
- addi rMEMP3,rMEMP,32
- beq L(nzCacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- addi rMEMP,rMEMP,32
- andi. rTMP,rMEMP3,127
- std rCHR,-8(rMEMP3)
-
- beq L(nzCacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP3)
- addi rMEMP,rMEMP,32
- std rCHR,8(rMEMP3)
- andi. rTMP,rMEMP,127
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
-
- beq L(nzCacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,32(rMEMP3)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,128
- std rCHR,40(rMEMP3)
- cmpldi cr6,rLEN,256
- li rMEMP2,128
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
- blt cr1,L(cacheAligned1)
- b L(nzCacheAligned128)
-
-/* Now we are aligned to the cache line and can use dcbtst. */
- .align 4
-L(nzCacheAligned):
- cmpldi cr1,rLEN,128
- blt cr1,L(cacheAligned1)
- b L(nzCacheAligned128)
- .align 5
-L(nzCacheAligned128):
- cmpldi cr1,rLEN,256
- addi rMEMP3,rMEMP,64
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- std rCHR,24(rMEMP)
- std rCHR,32(rMEMP)
- std rCHR,40(rMEMP)
- std rCHR,48(rMEMP)
- std rCHR,56(rMEMP)
- addi rMEMP,rMEMP3,64
- addi rLEN,rLEN,-128
- std rCHR,0(rMEMP3)
- std rCHR,8(rMEMP3)
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
- std rCHR,32(rMEMP3)
- std rCHR,40(rMEMP3)
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
- bge cr1,L(nzCacheAligned128)
- dcbtst 0,rMEMP
- b L(cacheAligned1)
- .align 5
-/* Storing a zero "c" value. We are aligned at a sector (32-byte)
- boundary but may not be at cache line (128-byte) boundary. If the
- remaining length spans a full cache line we can use the Data cache
- block zero instruction. */
-L(zloopstart):
-/* memset in 32-byte chunks until we get to a cache line boundary.
- If rLEN is less than the distance to the next cache-line boundary use
- cacheAligned1 code to finish the tail. */
- cmpldi cr1,rLEN,128
- beq L(medium)
-L(getCacheAligned):
- andi. rTMP,rMEMP,127
- nop
- blt cr1,L(cacheAligned1)
- addi rMEMP3,rMEMP,32
- beq L(cacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- addi rMEMP,rMEMP,32
- andi. rTMP,rMEMP3,127
- std rCHR,-8(rMEMP3)
-L(getCacheAligned2):
- beq L(cacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP3)
- std rCHR,8(rMEMP3)
- addi rMEMP,rMEMP,32
- andi. rTMP,rMEMP,127
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
-L(getCacheAligned3):
- beq L(cacheAligned)
- addi rLEN,rLEN,-32
- std rCHR,32(rMEMP3)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,128
- std rCHR,40(rMEMP3)
- cmpldi cr6,rLEN,256
- li rMEMP2,128
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
- blt cr1,L(cacheAligned1)
- blt cr6,L(cacheAligned128)
- b L(cacheAlignedx)
-
-/* Now we are aligned to the cache line and can use dcbz. */
- .align 5
-L(cacheAligned):
- cmpldi cr1,rLEN,128
- cmpldi cr6,rLEN,256
- blt cr1,L(cacheAligned1)
- li rMEMP2,128
-L(cacheAlignedx):
- cmpldi cr5,rLEN,640
- blt cr6,L(cacheAligned128)
- bgt cr5,L(cacheAligned512)
- cmpldi cr6,rLEN,512
- dcbz 0,rMEMP
- cmpldi cr1,rLEN,384
- dcbz rMEMP2,rMEMP
- addi rMEMP,rMEMP,256
- addi rLEN,rLEN,-256
- blt cr1,L(cacheAligned1)
- blt cr6,L(cacheAligned128)
- b L(cacheAligned256)
- .align 5
-/* A simple loop for the longer (>640 bytes) lengths. This form limits
- the branch miss-predicted to exactly 1 at loop exit.*/
-L(cacheAligned512):
- cmpldi cr1,rLEN,128
- blt cr1,L(cacheAligned1)
- dcbz 0,rMEMP
- addi rLEN,rLEN,-128
- addi rMEMP,rMEMP,128
- b L(cacheAligned512)
- .align 5
-L(cacheAligned256):
-
- cmpldi cr6,rLEN,512
-
- dcbz 0,rMEMP
- cmpldi cr1,rLEN,384
- dcbz rMEMP2,rMEMP
- addi rMEMP,rMEMP,256
- addi rLEN,rLEN,-256
-
- bge cr6,L(cacheAligned256)
-
- blt cr1,L(cacheAligned1)
- .align 4
-L(cacheAligned128):
- dcbz 0,rMEMP
- addi rMEMP,rMEMP,128
- addi rLEN,rLEN,-128
- nop
-L(cacheAligned1):
- cmpldi cr1,rLEN,32
- blt cr1,L(handletail32)
- addi rMEMP3,rMEMP,32
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP)
- std rCHR,8(rMEMP)
- std rCHR,16(rMEMP)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,32
- std rCHR,-8(rMEMP3)
-L(cacheAligned2):
- blt cr1,L(handletail32)
- addi rLEN,rLEN,-32
- std rCHR,0(rMEMP3)
- std rCHR,8(rMEMP3)
- addi rMEMP,rMEMP,32
- cmpldi cr1,rLEN,32
- std rCHR,16(rMEMP3)
- std rCHR,24(rMEMP3)
- nop
-L(cacheAligned3):
- blt cr1,L(handletail32)
- addi rMEMP,rMEMP,32
- addi rLEN,rLEN,-32
- std rCHR,32(rMEMP3)
- std rCHR,40(rMEMP3)
- std rCHR,48(rMEMP3)
- std rCHR,56(rMEMP3)
-
-/* We are here because the length or remainder (rLEN) is less than the
- cache line/sector size and does not justify aggressive loop unrolling.
- So set up the preconditions for L(medium) and go there. */
- .align 3
-L(handletail32):
- cmpldi cr1,rLEN,0
- beqlr cr1
- b L(medium)
-
- .align 5
-L(small):
-/* Memset of 8 bytes or less. */
- cmpldi cr6, rLEN, 4
- cmpldi cr5, rLEN, 1
- ble cr6,L(le4)
- subi rLEN, rLEN, 4
- stb rCHR,0(rMEMP)
- stb rCHR,1(rMEMP)
- stb rCHR,2(rMEMP)
- stb rCHR,3(rMEMP)
- addi rMEMP,rMEMP, 4
- cmpldi cr5, rLEN, 1
-L(le4):
- cmpldi cr1, rLEN, 3
- bltlr cr5
- stb rCHR, 0(rMEMP)
- beqlr cr5
- stb rCHR, 1(rMEMP)
- bltlr cr1
- stb rCHR, 2(rMEMP)
- beqlr cr1
- stb rCHR, 3(rMEMP)
- blr
-
-/* Memset of 0-31 bytes. */
- .align 5
-L(medium):
- insrdi rCHR, rCHR, 32, 0 /* Replicate word to double word. */
- cmpldi cr1, rLEN, 16
-L(medium_tail2):
- add rMEMP, rMEMP, rLEN
-L(medium_tail):
- bt- 31, L(medium_31t)
- bt- 30, L(medium_30t)
-L(medium_30f):
- bt 29, L(medium_29t)
-L(medium_29f):
- bge cr1, L(medium_27t)
- bflr 28
- std rCHR, -8(rMEMP)
- blr
-
-L(medium_31t):
- stbu rCHR, -1(rMEMP)
- bf- 30, L(medium_30f)
-L(medium_30t):
- sthu rCHR, -2(rMEMP)
- bf- 29, L(medium_29f)
-L(medium_29t):
- stwu rCHR, -4(rMEMP)
- blt cr1, L(medium_27f)
-L(medium_27t):
- std rCHR, -8(rMEMP)
- stdu rCHR, -16(rMEMP)
-L(medium_27f):
- bflr 28
-L(medium_28t):
- std rCHR, -8(rMEMP)
- blr
-END_GEN_TB (MEMSET,TB_TOCLESS)
-libc_hidden_builtin_def (memset)
-
-/* Copied from bzero.S to prevent the linker from inserting a stub
- between bzero and memset. */
-ENTRY (__bzero)
- CALL_MCOUNT 3
- mr r5,r4
- li r4,0
- b L(_memset)
-END (__bzero)
-#ifndef __bzero
-weak_alias (__bzero, bzero)
-#endif
diff --git a/sysdeps/powerpc/powerpc64/power6/multiarch/Implies b/sysdeps/powerpc/powerpc64/power6/multiarch/Implies
deleted file mode 100644
index 2ebe304fa6..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/multiarch/Implies
+++ /dev/null
@@ -1 +0,0 @@
-powerpc/powerpc64/power5+/multiarch
diff --git a/sysdeps/powerpc/powerpc64/power6/wcschr.c b/sysdeps/powerpc/powerpc64/power6/wcschr.c
deleted file mode 100644
index ae04a130cc..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/wcschr.c
+++ /dev/null
@@ -1 +0,0 @@
-#include <sysdeps/powerpc/power6/wcschr.c>
diff --git a/sysdeps/powerpc/powerpc64/power6/wcscpy.c b/sysdeps/powerpc/powerpc64/power6/wcscpy.c
deleted file mode 100644
index 722c8f995b..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/wcscpy.c
+++ /dev/null
@@ -1 +0,0 @@
-#include <sysdeps/powerpc/power6/wcscpy.c>
diff --git a/sysdeps/powerpc/powerpc64/power6/wcsrchr.c b/sysdeps/powerpc/powerpc64/power6/wcsrchr.c
deleted file mode 100644
index b86472d7bd..0000000000
--- a/sysdeps/powerpc/powerpc64/power6/wcsrchr.c
+++ /dev/null
@@ -1 +0,0 @@
-#include <sysdeps/powerpc/power6/wcsrchr.c>