076f82
commit df5de87260dba479873b2850bbe5c0b81c2376f6
076f82
Author: Noah Goldstein <goldstein.w.n@gmail.com>
076f82
Date:   Fri Apr 15 12:28:01 2022 -0500
076f82
076f82
    x86: Cleanup page cross code in memcmp-avx2-movbe.S
076f82
    
076f82
    Old code was both inefficient and wasted code size. New code (-62
076f82
    bytes) and comparable or better performance in the page cross case.
076f82
    
076f82
    geometric_mean(N=20) of page cross cases New / Original: 0.960
076f82
    
076f82
    size, align0, align1, ret, New Time/Old Time
076f82
       1,   4095,      0,   0,             1.001
076f82
       1,   4095,      0,   1,             0.999
076f82
       1,   4095,      0,  -1,               1.0
076f82
       2,   4094,      0,   0,               1.0
076f82
       2,   4094,      0,   1,               1.0
076f82
       2,   4094,      0,  -1,               1.0
076f82
       3,   4093,      0,   0,               1.0
076f82
       3,   4093,      0,   1,               1.0
076f82
       3,   4093,      0,  -1,               1.0
076f82
       4,   4092,      0,   0,             0.987
076f82
       4,   4092,      0,   1,               1.0
076f82
       4,   4092,      0,  -1,               1.0
076f82
       5,   4091,      0,   0,             0.984
076f82
       5,   4091,      0,   1,             1.002
076f82
       5,   4091,      0,  -1,             1.005
076f82
       6,   4090,      0,   0,             0.993
076f82
       6,   4090,      0,   1,             1.001
076f82
       6,   4090,      0,  -1,             1.003
076f82
       7,   4089,      0,   0,             0.991
076f82
       7,   4089,      0,   1,               1.0
076f82
       7,   4089,      0,  -1,             1.001
076f82
       8,   4088,      0,   0,             0.875
076f82
       8,   4088,      0,   1,             0.881
076f82
       8,   4088,      0,  -1,             0.888
076f82
       9,   4087,      0,   0,             0.872
076f82
       9,   4087,      0,   1,             0.879
076f82
       9,   4087,      0,  -1,             0.883
076f82
      10,   4086,      0,   0,             0.878
076f82
      10,   4086,      0,   1,             0.886
076f82
      10,   4086,      0,  -1,             0.873
076f82
      11,   4085,      0,   0,             0.878
076f82
      11,   4085,      0,   1,             0.881
076f82
      11,   4085,      0,  -1,             0.879
076f82
      12,   4084,      0,   0,             0.873
076f82
      12,   4084,      0,   1,             0.889
076f82
      12,   4084,      0,  -1,             0.875
076f82
      13,   4083,      0,   0,             0.873
076f82
      13,   4083,      0,   1,             0.863
076f82
      13,   4083,      0,  -1,             0.863
076f82
      14,   4082,      0,   0,             0.838
076f82
      14,   4082,      0,   1,             0.869
076f82
      14,   4082,      0,  -1,             0.877
076f82
      15,   4081,      0,   0,             0.841
076f82
      15,   4081,      0,   1,             0.869
076f82
      15,   4081,      0,  -1,             0.876
076f82
      16,   4080,      0,   0,             0.988
076f82
      16,   4080,      0,   1,              0.99
076f82
      16,   4080,      0,  -1,             0.989
076f82
      17,   4079,      0,   0,             0.978
076f82
      17,   4079,      0,   1,             0.981
076f82
      17,   4079,      0,  -1,              0.98
076f82
      18,   4078,      0,   0,             0.981
076f82
      18,   4078,      0,   1,              0.98
076f82
      18,   4078,      0,  -1,             0.985
076f82
      19,   4077,      0,   0,             0.977
076f82
      19,   4077,      0,   1,             0.979
076f82
      19,   4077,      0,  -1,             0.986
076f82
      20,   4076,      0,   0,             0.977
076f82
      20,   4076,      0,   1,             0.986
076f82
      20,   4076,      0,  -1,             0.984
076f82
      21,   4075,      0,   0,             0.977
076f82
      21,   4075,      0,   1,             0.983
076f82
      21,   4075,      0,  -1,             0.988
076f82
      22,   4074,      0,   0,             0.983
076f82
      22,   4074,      0,   1,             0.994
076f82
      22,   4074,      0,  -1,             0.993
076f82
      23,   4073,      0,   0,              0.98
076f82
      23,   4073,      0,   1,             0.992
076f82
      23,   4073,      0,  -1,             0.995
076f82
      24,   4072,      0,   0,             0.989
076f82
      24,   4072,      0,   1,             0.989
076f82
      24,   4072,      0,  -1,             0.991
076f82
      25,   4071,      0,   0,              0.99
076f82
      25,   4071,      0,   1,             0.999
076f82
      25,   4071,      0,  -1,             0.996
076f82
      26,   4070,      0,   0,             0.993
076f82
      26,   4070,      0,   1,             0.995
076f82
      26,   4070,      0,  -1,             0.998
076f82
      27,   4069,      0,   0,             0.993
076f82
      27,   4069,      0,   1,             0.999
076f82
      27,   4069,      0,  -1,               1.0
076f82
      28,   4068,      0,   0,             0.997
076f82
      28,   4068,      0,   1,               1.0
076f82
      28,   4068,      0,  -1,             0.999
076f82
      29,   4067,      0,   0,             0.996
076f82
      29,   4067,      0,   1,             0.999
076f82
      29,   4067,      0,  -1,             0.999
076f82
      30,   4066,      0,   0,             0.991
076f82
      30,   4066,      0,   1,             1.001
076f82
      30,   4066,      0,  -1,             0.999
076f82
      31,   4065,      0,   0,             0.988
076f82
      31,   4065,      0,   1,             0.998
076f82
      31,   4065,      0,  -1,             0.998
076f82
    Reviewed-by: H.J. Lu <hjl.tools@gmail.com>
076f82
    
076f82
    (cherry picked from commit 23102686ec67b856a2d4fd25ddaa1c0b8d175c4f)
076f82
076f82
diff --git a/sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S b/sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S
076f82
index 2621ec907aedb781..ec9cf0852edf216d 100644
076f82
--- a/sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S
076f82
+++ b/sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S
076f82
@@ -429,22 +429,21 @@ L(page_cross_less_vec):
076f82
 # ifndef USE_AS_WMEMCMP
076f82
 	cmpl	$8, %edx
076f82
 	jae	L(between_8_15)
076f82
+	/* Fall through for [4, 7].  */
076f82
 	cmpl	$4, %edx
076f82
-	jae	L(between_4_7)
076f82
+	jb	L(between_2_3)
076f82
 
076f82
-	/* Load as big endian to avoid branches.  */
076f82
-	movzwl	(%rdi), %eax
076f82
-	movzwl	(%rsi), %ecx
076f82
-	shll	$8, %eax
076f82
-	shll	$8, %ecx
076f82
-	bswap	%eax
076f82
-	bswap	%ecx
076f82
-	movzbl	-1(%rdi, %rdx), %edi
076f82
-	movzbl	-1(%rsi, %rdx), %esi
076f82
-	orl	%edi, %eax
076f82
-	orl	%esi, %ecx
076f82
-	/* Subtraction is okay because the upper 8 bits are zero.  */
076f82
-	subl	%ecx, %eax
076f82
+	movbe	(%rdi), %eax
076f82
+	movbe	(%rsi), %ecx
076f82
+	shlq	$32, %rax
076f82
+	shlq	$32, %rcx
076f82
+	movbe	-4(%rdi, %rdx), %edi
076f82
+	movbe	-4(%rsi, %rdx), %esi
076f82
+	orq	%rdi, %rax
076f82
+	orq	%rsi, %rcx
076f82
+	subq	%rcx, %rax
076f82
+	/* Fast path for return zero.  */
076f82
+	jnz	L(ret_nonzero)
076f82
 	/* No ymm register was touched.  */
076f82
 	ret
076f82
 
076f82
@@ -457,9 +456,33 @@ L(one_or_less):
076f82
 	/* No ymm register was touched.  */
076f82
 	ret
076f82
 
076f82
+	.p2align 4,, 5
076f82
+L(ret_nonzero):
076f82
+	sbbl	%eax, %eax
076f82
+	orl	$1, %eax
076f82
+	/* No ymm register was touched.  */
076f82
+	ret
076f82
+
076f82
+	.p2align 4,, 2
076f82
+L(zero):
076f82
+	xorl	%eax, %eax
076f82
+	/* No ymm register was touched.  */
076f82
+	ret
076f82
+
076f82
 	.p2align 4
076f82
 L(between_8_15):
076f82
-# endif
076f82
+	movbe	(%rdi), %rax
076f82
+	movbe	(%rsi), %rcx
076f82
+	subq	%rcx, %rax
076f82
+	jnz	L(ret_nonzero)
076f82
+	movbe	-8(%rdi, %rdx), %rax
076f82
+	movbe	-8(%rsi, %rdx), %rcx
076f82
+	subq	%rcx, %rax
076f82
+	/* Fast path for return zero.  */
076f82
+	jnz	L(ret_nonzero)
076f82
+	/* No ymm register was touched.  */
076f82
+	ret
076f82
+# else
076f82
 	/* If USE_AS_WMEMCMP fall through into 8-15 byte case.  */
076f82
 	vmovq	(%rdi), %xmm1
076f82
 	vmovq	(%rsi), %xmm2
076f82
@@ -475,16 +498,13 @@ L(between_8_15):
076f82
 	VPCMPEQ	%xmm1, %xmm2, %xmm2
076f82
 	vpmovmskb %xmm2, %eax
076f82
 	subl	$0xffff, %eax
076f82
+	/* Fast path for return zero.  */
076f82
 	jnz	L(return_vec_0)
076f82
 	/* No ymm register was touched.  */
076f82
 	ret
076f82
+# endif
076f82
 
076f82
-	.p2align 4
076f82
-L(zero):
076f82
-	xorl	%eax, %eax
076f82
-	ret
076f82
-
076f82
-	.p2align 4
076f82
+	.p2align 4,, 10
076f82
 L(between_16_31):
076f82
 	/* From 16 to 31 bytes.  No branch when size == 16.  */
076f82
 	vmovdqu	(%rsi), %xmm2
076f82
@@ -501,11 +521,17 @@ L(between_16_31):
076f82
 	VPCMPEQ	(%rdi), %xmm2, %xmm2
076f82
 	vpmovmskb %xmm2, %eax
076f82
 	subl	$0xffff, %eax
076f82
+	/* Fast path for return zero.  */
076f82
 	jnz	L(return_vec_0)
076f82
 	/* No ymm register was touched.  */
076f82
 	ret
076f82
 
076f82
 # ifdef USE_AS_WMEMCMP
076f82
+	.p2align 4,, 2
076f82
+L(zero):
076f82
+	xorl	%eax, %eax
076f82
+	ret
076f82
+
076f82
 	.p2align 4
076f82
 L(one_or_less):
076f82
 	jb	L(zero)
076f82
@@ -520,22 +546,20 @@ L(one_or_less):
076f82
 # else
076f82
 
076f82
 	.p2align 4
076f82
-L(between_4_7):
076f82
-	/* Load as big endian with overlapping movbe to avoid branches.
076f82
-	 */
076f82
-	movbe	(%rdi), %eax
076f82
-	movbe	(%rsi), %ecx
076f82
-	shlq	$32, %rax
076f82
-	shlq	$32, %rcx
076f82
-	movbe	-4(%rdi, %rdx), %edi
076f82
-	movbe	-4(%rsi, %rdx), %esi
076f82
-	orq	%rdi, %rax
076f82
-	orq	%rsi, %rcx
076f82
-	subq	%rcx, %rax
076f82
-	jz	L(zero_4_7)
076f82
-	sbbl	%eax, %eax
076f82
-	orl	$1, %eax
076f82
-L(zero_4_7):
076f82
+L(between_2_3):
076f82
+	/* Load as big endian to avoid branches.  */
076f82
+	movzwl	(%rdi), %eax
076f82
+	movzwl	(%rsi), %ecx
076f82
+	bswap	%eax
076f82
+	bswap	%ecx
076f82
+	shrl	%eax
076f82
+	shrl	%ecx
076f82
+	movzbl	-1(%rdi, %rdx), %edi
076f82
+	movzbl	-1(%rsi, %rdx), %esi
076f82
+	orl	%edi, %eax
076f82
+	orl	%esi, %ecx
076f82
+	/* Subtraction is okay because the upper bit is zero.  */
076f82
+	subl	%ecx, %eax
076f82
 	/* No ymm register was touched.  */
076f82
 	ret
076f82
 # endif