aboutsummaryrefslogtreecommitdiff
path: root/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
diff options
context:
space:
mode:
authorH.J. Lu <hjl.tools@gmail.com>2016-06-30 07:57:07 -0700
committerH.J. Lu <hjl.tools@gmail.com>2016-06-30 07:58:11 -0700
commit13efa86ece61bf84daca50cab30db1b0902fe2db (patch)
treee6ed4e21bb720ba21d069d61064a057be6f999db /sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
parent73fb56a4d51fd4437e4cde6dd3c8077a610f88a8 (diff)
downloadglibc-13efa86ece61bf84daca50cab30db1b0902fe2db.tar
glibc-13efa86ece61bf84daca50cab30db1b0902fe2db.tar.gz
glibc-13efa86ece61bf84daca50cab30db1b0902fe2db.tar.bz2
glibc-13efa86ece61bf84daca50cab30db1b0902fe2db.zip
Check Prefer_ERMS in memmove/memcpy/mempcpy/memset
Although the Enhanced REP MOVSB/STOSB (ERMS) implementations of memmove, memcpy, mempcpy and memset aren't used by the current processors, this patch adds Prefer_ERMS check in memmove, memcpy, mempcpy and memset so that they can be used in the future. * sysdeps/x86/cpu-features.h (bit_arch_Prefer_ERMS): New. (index_arch_Prefer_ERMS): Likewise. * sysdeps/x86_64/multiarch/memcpy.S (__new_memcpy): Return __memcpy_erms for Prefer_ERMS. * sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S (__memmove_erms): Enabled for libc.a. * ysdeps/x86_64/multiarch/memmove.S (__libc_memmove): Return __memmove_erms or Prefer_ERMS. * sysdeps/x86_64/multiarch/mempcpy.S (__mempcpy): Return __mempcpy_erms for Prefer_ERMS. * sysdeps/x86_64/multiarch/memset.S (memset): Return __memset_erms for Prefer_ERMS.
Diffstat (limited to 'sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S')
-rw-r--r--sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S6
1 files changed, 5 insertions, 1 deletions
diff --git a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
index a2cce39a16..4893ea46b4 100644
--- a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
+++ b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
@@ -150,13 +150,15 @@ L(nop):
#if defined USE_MULTIARCH && IS_IN (libc)
END (MEMMOVE_SYMBOL (__memmove, unaligned))
-# if VEC_SIZE == 16 && defined SHARED
+# if VEC_SIZE == 16
+# if defined SHARED
/* Only used to measure performance of REP MOVSB. */
ENTRY (__mempcpy_erms)
movq %rdi, %rax
addq %rdx, %rax
jmp L(start_movsb)
END (__mempcpy_erms)
+# endif
ENTRY (__memmove_erms)
movq %rdi, %rax
@@ -181,7 +183,9 @@ L(movsb_backward):
cld
ret
END (__memmove_erms)
+# if defined SHARED
strong_alias (__memmove_erms, __memcpy_erms)
+# endif
# endif
# ifdef SHARED