aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--sysdeps/powerpc/powerpc32/cell/memcpy.S24
-rw-r--r--sysdeps/powerpc/powerpc64/cell/memcpy.S24
2 files changed, 24 insertions, 24 deletions
diff --git a/sysdeps/powerpc/powerpc32/cell/memcpy.S b/sysdeps/powerpc/powerpc32/cell/memcpy.S
index e6c076cbe1..cc1da99fd9 100644
--- a/sysdeps/powerpc/powerpc32/cell/memcpy.S
+++ b/sysdeps/powerpc/powerpc32/cell/memcpy.S
@@ -43,16 +43,16 @@
.align 7
EALIGN (BP_SYM (memcpy), 5, 0)
- CALL_MCOUNT
+ CALL_MCOUNT
dcbt 0,r4 /* Prefetch ONE SRC cacheline */
cmplwi cr1,r5,16 /* is size < 16 ? */
- mr r6,r3
+ mr r6,r3
blt+ cr1,.Lshortcopy
.Lbigcopy:
neg r8,r3 /* LS 3 bits = # bytes to 8-byte dest bdry */
- clrlwi r8,r8,32-4 /* aling to 16byte boundary */
+ clrlwi r8,r8,32-4 /* aling to 16byte boundary */
sub r7,r4,r3
cmplwi cr0,r8,0
beq+ .Ldst_aligned
@@ -112,8 +112,8 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.LprefetchSRC:
dcbt r12,r4
- addi r12,r12,128
- bdnz .LprefetchSRC
+ addi r12,r12,128
+ bdnz .LprefetchSRC
.Lnocacheprefetch:
mtctr r7
@@ -122,7 +122,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
beq cr6,.Lcachelinealigned
.Laligntocacheline:
- lfd fp9,0x08(r4)
+ lfd fp9,0x08(r4)
lfdu fp10,0x10(r4)
stfd fp9,0x08(r6)
stfdu fp10,0x10(r6)
@@ -131,10 +131,10 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.Lcachelinealigned: /* copy while cache lines */
- blt- cr1,.Llessthancacheline /* size <128 */
+ blt- cr1,.Llessthancacheline /* size <128 */
.Louterloop:
- cmpwi r11,0
+ cmpwi r11,0
mtctr r11
beq- .Lendloop
@@ -142,7 +142,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.align 4
/* Copy whole cachelines, optimized by prefetching SRC cacheline */
-.Lloop: /* Copy aligned body */
+.Lloop: /* Copy aligned body */
dcbt r12,r4 /* PREFETCH SOURCE some cache lines ahead */
lfd fp9, 0x08(r4)
dcbz r11,r6
@@ -186,7 +186,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
beq- .Lendloop2
mtctr r10
-.Lloop2: /* Copy aligned body */
+.Lloop2: /* Copy aligned body */
lfd fp9, 0x08(r4)
lfd fp10, 0x10(r4)
lfd fp11, 0x18(r4)
@@ -206,7 +206,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
mtctr r7
.Lcopy_remaining:
- lfd fp9,0x08(r4)
+ lfd fp9,0x08(r4)
lfdu fp10,0x10(r4)
stfd fp9,0x08(r6)
stfdu fp10,0x10(r6)
@@ -214,7 +214,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.Ldo_lt16: /* less than 16 ? */
cmplwi cr0,r5,0 /* copy remaining bytes (0-15) */
- beqlr+ /* no rest to copy */
+ beqlr+ /* no rest to copy */
addi r4,r4,8
addi r6,r6,8
diff --git a/sysdeps/powerpc/powerpc64/cell/memcpy.S b/sysdeps/powerpc/powerpc64/cell/memcpy.S
index 2a00a6ed52..c6ee730e4e 100644
--- a/sysdeps/powerpc/powerpc64/cell/memcpy.S
+++ b/sysdeps/powerpc/powerpc64/cell/memcpy.S
@@ -43,16 +43,16 @@
.align 7
EALIGN (BP_SYM (memcpy), 5, 0)
- CALL_MCOUNT 3
+ CALL_MCOUNT 3
dcbt 0,r4 /* Prefetch ONE SRC cacheline */
cmpldi cr1,r5,16 /* is size < 16 ? */
- mr r6,r3
+ mr r6,r3
blt+ cr1,.Lshortcopy
.Lbigcopy:
neg r8,r3 /* LS 3 bits = # bytes to 8-byte dest bdry */
- clrldi r8,r8,64-4 /* aling to 16byte boundary */
+ clrldi r8,r8,64-4 /* aling to 16byte boundary */
sub r7,r4,r3
cmpldi cr0,r8,0
beq+ .Ldst_aligned
@@ -112,8 +112,8 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.LprefetchSRC:
dcbt r12,r4
- addi r12,r12,128
- bdnz .LprefetchSRC
+ addi r12,r12,128
+ bdnz .LprefetchSRC
.Lnocacheprefetch:
mtctr r7
@@ -122,7 +122,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
beq cr6,.Lcachelinealigned
.Laligntocacheline:
- ld r9,0x08(r4)
+ ld r9,0x08(r4)
ldu r7,0x10(r4)
std r9,0x08(r6)
stdu r7,0x10(r6)
@@ -131,10 +131,10 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.Lcachelinealigned: /* copy while cache lines */
- blt- cr1,.Llessthancacheline /* size <128 */
+ blt- cr1,.Llessthancacheline /* size <128 */
.Louterloop:
- cmpdi r11,0
+ cmpdi r11,0
mtctr r11
beq- .Lendloop
@@ -142,7 +142,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.align 4
/* Copy whole cachelines, optimized by prefetching SRC cacheline */
-.Lloop: /* Copy aligned body */
+.Lloop: /* Copy aligned body */
dcbt r12,r4 /* PREFETCH SOURCE some cache lines ahead */
ld r9, 0x08(r4)
dcbz r11,r6
@@ -186,7 +186,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
beq- .Lendloop2
mtctr r10
-.Lloop2: /* Copy aligned body */
+.Lloop2: /* Copy aligned body */
ld r9, 0x08(r4)
ld r7, 0x10(r4)
ld r8, 0x18(r4)
@@ -206,7 +206,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
mtctr r7
.Lcopy_remaining:
- ld r8,0x08(r4)
+ ld r8,0x08(r4)
ldu r7,0x10(r4)
std r8,0x08(r6)
stdu r7,0x10(r6)
@@ -214,7 +214,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
.Ldo_lt16: /* less than 16 ? */
cmpldi cr0,r5,0 /* copy remaining bytes (0-15) */
- beqlr+ /* no rest to copy */
+ beqlr+ /* no rest to copy */
addi r4,r4,8
addi r6,r6,8