forked from rpms/glibc
668eaab0c7
* Fri Jul 22 2022 Arjun Shankar <arjun@redhat.com> - 2.34-40 - Sync with upstream branch release/2.34/master, commit b2f32e746492615a6eb3e66fac1e766e32e8deb1: - malloc: Simplify implementation of __malloc_assert - Update syscall-names.list for Linux 5.18 - x86: Add missing IS_IN (libc) check to strncmp-sse4_2.S - x86: Move mem{p}{mov|cpy}_{chk_}erms to its own file - x86: Move and slightly improve memset_erms - x86: Add definition for __wmemset_chk AVX2 RTM in ifunc impl list - x86: Put wcs{n}len-sse4.1 in the sse4.1 text section - x86: Align entry for memrchr to 64-bytes. - x86: Add BMI1/BMI2 checks for ISA_V3 check - x86: Cleanup bounds checking in large memcpy case - x86: Add bounds `x86_non_temporal_threshold` - x86: Add sse42 implementation to strcmp's ifunc - x86: Fix misordered logic for setting `rep_movsb_stop_threshold` - x86: Align varshift table to 32-bytes - x86: ZERO_UPPER_VEC_REGISTERS_RETURN_XTEST expect no transactions - x86: Shrink code size of memchr-evex.S - x86: Shrink code size of memchr-avx2.S - x86: Optimize memrchr-avx2.S - x86: Optimize memrchr-evex.S - x86: Optimize memrchr-sse2.S - x86: Add COND_VZEROUPPER that can replace vzeroupper if no `ret` - x86: Create header for VEC classes in x86 strings library - x86_64: Add strstr function with 512-bit EVEX - x86-64: Ignore r_addend for R_X86_64_GLOB_DAT/R_X86_64_JUMP_SLOT - x86_64: Implement evex512 version of strlen, strnlen, wcslen and wcsnlen - x86_64: Remove bzero optimization - x86_64: Remove end of line trailing spaces - nptl: Fix ___pthread_unregister_cancel_restore asynchronous restore - linux: Fix mq_timereceive check for 32 bit fallback code (BZ 29304) Resolves: #2109505
89 lines
3.3 KiB
Diff
89 lines
3.3 KiB
Diff
commit ba1c3f23d9ba63c38333116eec6043c471c378c4
|
|
Author: Noah Goldstein <goldstein.w.n@gmail.com>
|
|
Date: Wed Jun 15 10:41:28 2022 -0700
|
|
|
|
x86: Cleanup bounds checking in large memcpy case
|
|
|
|
1. Fix incorrect lower-bound threshold in L(large_memcpy_2x).
|
|
Previously was using `__x86_rep_movsb_threshold` and should
|
|
have been using `__x86_shared_non_temporal_threshold`.
|
|
|
|
2. Avoid reloading __x86_shared_non_temporal_threshold before
|
|
the L(large_memcpy_4x) bounds check.
|
|
|
|
3. Document the second bounds check for L(large_memcpy_4x)
|
|
more clearly.
|
|
|
|
(cherry picked from commit 89a25c6f64746732b87eaf433af0964b564d4a92)
|
|
|
|
diff --git a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
|
|
index 7b27cbdda5fb99f7..618d46d8ce28828c 100644
|
|
--- a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
|
|
+++ b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
|
|
@@ -118,7 +118,13 @@
|
|
# define LARGE_LOAD_SIZE (VEC_SIZE * 4)
|
|
#endif
|
|
|
|
-/* Amount to shift rdx by to compare for memcpy_large_4x. */
|
|
+/* Amount to shift __x86_shared_non_temporal_threshold by for
|
|
+ bound for memcpy_large_4x. This is essentially use to to
|
|
+ indicate that the copy is far beyond the scope of L3
|
|
+ (assuming no user config x86_non_temporal_threshold) and to
|
|
+ use a more aggressively unrolled loop. NB: before
|
|
+ increasing the value also update initialization of
|
|
+ x86_non_temporal_threshold. */
|
|
#ifndef LOG_4X_MEMCPY_THRESH
|
|
# define LOG_4X_MEMCPY_THRESH 4
|
|
#endif
|
|
@@ -724,9 +730,14 @@ L(skip_short_movsb_check):
|
|
.p2align 4,, 10
|
|
#if (defined USE_MULTIARCH || VEC_SIZE == 16) && IS_IN (libc)
|
|
L(large_memcpy_2x_check):
|
|
- cmp __x86_rep_movsb_threshold(%rip), %RDX_LP
|
|
- jb L(more_8x_vec_check)
|
|
+ /* Entry from L(large_memcpy_2x) has a redundant load of
|
|
+ __x86_shared_non_temporal_threshold(%rip). L(large_memcpy_2x)
|
|
+ is only use for the non-erms memmove which is generally less
|
|
+ common. */
|
|
L(large_memcpy_2x):
|
|
+ mov __x86_shared_non_temporal_threshold(%rip), %R11_LP
|
|
+ cmp %R11_LP, %RDX_LP
|
|
+ jb L(more_8x_vec_check)
|
|
/* To reach this point it is impossible for dst > src and
|
|
overlap. Remaining to check is src > dst and overlap. rcx
|
|
already contains dst - src. Negate rcx to get src - dst. If
|
|
@@ -774,18 +785,21 @@ L(large_memcpy_2x):
|
|
/* ecx contains -(dst - src). not ecx will return dst - src - 1
|
|
which works for testing aliasing. */
|
|
notl %ecx
|
|
+ movq %rdx, %r10
|
|
testl $(PAGE_SIZE - VEC_SIZE * 8), %ecx
|
|
jz L(large_memcpy_4x)
|
|
|
|
- movq %rdx, %r10
|
|
- shrq $LOG_4X_MEMCPY_THRESH, %r10
|
|
- cmp __x86_shared_non_temporal_threshold(%rip), %r10
|
|
+ /* r11 has __x86_shared_non_temporal_threshold. Shift it left
|
|
+ by LOG_4X_MEMCPY_THRESH to get L(large_memcpy_4x) threshold.
|
|
+ */
|
|
+ shlq $LOG_4X_MEMCPY_THRESH, %r11
|
|
+ cmp %r11, %rdx
|
|
jae L(large_memcpy_4x)
|
|
|
|
/* edx will store remainder size for copying tail. */
|
|
andl $(PAGE_SIZE * 2 - 1), %edx
|
|
/* r10 stores outer loop counter. */
|
|
- shrq $((LOG_PAGE_SIZE + 1) - LOG_4X_MEMCPY_THRESH), %r10
|
|
+ shrq $(LOG_PAGE_SIZE + 1), %r10
|
|
/* Copy 4x VEC at a time from 2 pages. */
|
|
.p2align 4
|
|
L(loop_large_memcpy_2x_outer):
|
|
@@ -850,7 +864,6 @@ L(large_memcpy_2x_end):
|
|
|
|
.p2align 4
|
|
L(large_memcpy_4x):
|
|
- movq %rdx, %r10
|
|
/* edx will store remainder size for copying tail. */
|
|
andl $(PAGE_SIZE * 4 - 1), %edx
|
|
/* r10 stores outer loop counter. */
|