076f82
commit 5cb6329652696e79d6d576165ea87e332c9de106
076f82
Author: H.J. Lu <hjl.tools@gmail.com>
076f82
Date:   Mon Feb 7 05:55:15 2022 -0800
076f82
076f82
    x86-64: Optimize bzero
076f82
    
076f82
    memset with zero as the value to set is by far the majority value (99%+
076f82
    for Python3 and GCC).
076f82
    
076f82
    bzero can be slightly more optimized for this case by using a zero-idiom
076f82
    xor for broadcasting the set value to a register (vector or GPR).
076f82
    
076f82
    Co-developed-by: Noah Goldstein <goldstein.w.n@gmail.com>
076f82
    (cherry picked from commit 3d9f171bfb5325bd5f427e9fc386453358c6e840)
076f82
076f82
diff --git a/sysdeps/x86_64/memset.S b/sysdeps/x86_64/memset.S
076f82
index 954471e5a5bf225b..0358210c7ff3a976 100644
076f82
--- a/sysdeps/x86_64/memset.S
076f82
+++ b/sysdeps/x86_64/memset.S
076f82
@@ -35,6 +35,9 @@
076f82
   punpcklwd %xmm0, %xmm0; \
076f82
   pshufd $0, %xmm0, %xmm0
076f82
 
076f82
+# define BZERO_ZERO_VEC0() \
076f82
+  pxor %xmm0, %xmm0
076f82
+
076f82
 # define WMEMSET_SET_VEC0_AND_SET_RETURN(d, r) \
076f82
   movd d, %xmm0; \
076f82
   pshufd $0, %xmm0, %xmm0; \
076f82
@@ -53,6 +56,10 @@
076f82
 # define MEMSET_SYMBOL(p,s)	memset
076f82
 #endif
076f82
 
076f82
+#ifndef BZERO_SYMBOL
076f82
+# define BZERO_SYMBOL(p,s)	__bzero
076f82
+#endif
076f82
+
076f82
 #ifndef WMEMSET_SYMBOL
076f82
 # define WMEMSET_CHK_SYMBOL(p,s) p
076f82
 # define WMEMSET_SYMBOL(p,s)	__wmemset
076f82
@@ -63,6 +70,7 @@
076f82
 libc_hidden_builtin_def (memset)
076f82
 
076f82
 #if IS_IN (libc)
076f82
+weak_alias (__bzero, bzero)
076f82
 libc_hidden_def (__wmemset)
076f82
 weak_alias (__wmemset, wmemset)
076f82
 libc_hidden_weak (wmemset)
076f82
diff --git a/sysdeps/x86_64/multiarch/Makefile b/sysdeps/x86_64/multiarch/Makefile
076f82
index 26be40959ce62895..37d8d6f0bd2d10cc 100644
076f82
--- a/sysdeps/x86_64/multiarch/Makefile
076f82
+++ b/sysdeps/x86_64/multiarch/Makefile
076f82
@@ -1,85 +1,130 @@
076f82
 ifeq ($(subdir),string)
076f82
 
076f82
-sysdep_routines += strncat-c stpncpy-c strncpy-c \
076f82
-		   strcmp-sse2 strcmp-sse2-unaligned strcmp-ssse3  \
076f82
-		   strcmp-sse4_2 strcmp-avx2 \
076f82
-		   strncmp-sse2 strncmp-ssse3 strncmp-sse4_2 strncmp-avx2 \
076f82
-		   memchr-sse2 rawmemchr-sse2 memchr-avx2 rawmemchr-avx2 \
076f82
-		   memrchr-sse2 memrchr-avx2 \
076f82
-		   memcmp-sse2 \
076f82
-		   memcmp-avx2-movbe \
076f82
-		   memcmp-sse4 memcpy-ssse3 \
076f82
-		   memmove-ssse3 \
076f82
-		   memcpy-ssse3-back \
076f82
-		   memmove-ssse3-back \
076f82
-		   memmove-avx512-no-vzeroupper \
076f82
-		   strcasecmp_l-sse2 strcasecmp_l-ssse3 \
076f82
-		   strcasecmp_l-sse4_2 strcasecmp_l-avx \
076f82
-		   strncase_l-sse2 strncase_l-ssse3 \
076f82
-		   strncase_l-sse4_2 strncase_l-avx \
076f82
-		   strchr-sse2 strchrnul-sse2 strchr-avx2 strchrnul-avx2 \
076f82
-		   strrchr-sse2 strrchr-avx2 \
076f82
-		   strlen-sse2 strnlen-sse2 strlen-avx2 strnlen-avx2 \
076f82
-		   strcat-avx2 strncat-avx2 \
076f82
-		   strcat-ssse3 strncat-ssse3\
076f82
-		   strcpy-avx2 strncpy-avx2 \
076f82
-		   strcpy-sse2 stpcpy-sse2 \
076f82
-		   strcpy-ssse3 strncpy-ssse3 stpcpy-ssse3 stpncpy-ssse3 \
076f82
-		   strcpy-sse2-unaligned strncpy-sse2-unaligned \
076f82
-		   stpcpy-sse2-unaligned stpncpy-sse2-unaligned \
076f82
-		   stpcpy-avx2 stpncpy-avx2 \
076f82
-		   strcat-sse2 \
076f82
-		   strcat-sse2-unaligned strncat-sse2-unaligned \
076f82
-		   strchr-sse2-no-bsf memcmp-ssse3 strstr-sse2-unaligned \
076f82
-		   strcspn-sse2 strpbrk-sse2 strspn-sse2 \
076f82
-		   strcspn-c strpbrk-c strspn-c varshift \
076f82
-		   memset-avx512-no-vzeroupper \
076f82
-		   memmove-sse2-unaligned-erms \
076f82
-		   memmove-avx-unaligned-erms \
076f82
-		   memmove-avx512-unaligned-erms \
076f82
-		   memset-sse2-unaligned-erms \
076f82
-		   memset-avx2-unaligned-erms \
076f82
-		   memset-avx512-unaligned-erms \
076f82
-		   memchr-avx2-rtm \
076f82
-		   memcmp-avx2-movbe-rtm \
076f82
-		   memmove-avx-unaligned-erms-rtm \
076f82
-		   memrchr-avx2-rtm \
076f82
-		   memset-avx2-unaligned-erms-rtm \
076f82
-		   rawmemchr-avx2-rtm \
076f82
-		   strchr-avx2-rtm \
076f82
-		   strcmp-avx2-rtm \
076f82
-		   strchrnul-avx2-rtm \
076f82
-		   stpcpy-avx2-rtm \
076f82
-		   stpncpy-avx2-rtm \
076f82
-		   strcat-avx2-rtm \
076f82
-		   strcpy-avx2-rtm \
076f82
-		   strlen-avx2-rtm \
076f82
-		   strncat-avx2-rtm \
076f82
-		   strncmp-avx2-rtm \
076f82
-		   strncpy-avx2-rtm \
076f82
-		   strnlen-avx2-rtm \
076f82
-		   strrchr-avx2-rtm \
076f82
-		   memchr-evex \
076f82
-		   memcmp-evex-movbe \
076f82
-		   memmove-evex-unaligned-erms \
076f82
-		   memrchr-evex \
076f82
-		   memset-evex-unaligned-erms \
076f82
-		   rawmemchr-evex \
076f82
-		   stpcpy-evex \
076f82
-		   stpncpy-evex \
076f82
-		   strcat-evex \
076f82
-		   strchr-evex \
076f82
-		   strchrnul-evex \
076f82
-		   strcmp-evex \
076f82
-		   strcpy-evex \
076f82
-		   strlen-evex \
076f82
-		   strncat-evex \
076f82
-		   strncmp-evex \
076f82
-		   strncpy-evex \
076f82
-		   strnlen-evex \
076f82
-		   strrchr-evex \
076f82
-		   memchr-evex-rtm \
076f82
-		   rawmemchr-evex-rtm
076f82
+sysdep_routines += \
076f82
+  bzero \
076f82
+  memchr-avx2 \
076f82
+  memchr-avx2-rtm \
076f82
+  memchr-evex \
076f82
+  memchr-evex-rtm \
076f82
+  memchr-sse2 \
076f82
+  memcmp-avx2-movbe \
076f82
+  memcmp-avx2-movbe-rtm \
076f82
+  memcmp-evex-movbe \
076f82
+  memcmp-sse2 \
076f82
+  memcmp-sse4 \
076f82
+  memcmp-ssse3 \
076f82
+  memcpy-ssse3 \
076f82
+  memcpy-ssse3-back \
076f82
+  memmove-avx-unaligned-erms \
076f82
+  memmove-avx-unaligned-erms-rtm \
076f82
+  memmove-avx512-no-vzeroupper \
076f82
+  memmove-avx512-unaligned-erms \
076f82
+  memmove-evex-unaligned-erms \
076f82
+  memmove-sse2-unaligned-erms \
076f82
+  memmove-ssse3 \
076f82
+  memmove-ssse3-back \
076f82
+  memrchr-avx2 \
076f82
+  memrchr-avx2-rtm \
076f82
+  memrchr-evex \
076f82
+  memrchr-sse2 \
076f82
+  memset-avx2-unaligned-erms \
076f82
+  memset-avx2-unaligned-erms-rtm \
076f82
+  memset-avx512-no-vzeroupper \
076f82
+  memset-avx512-unaligned-erms \
076f82
+  memset-evex-unaligned-erms \
076f82
+  memset-sse2-unaligned-erms \
076f82
+  rawmemchr-avx2 \
076f82
+  rawmemchr-avx2-rtm \
076f82
+  rawmemchr-evex \
076f82
+  rawmemchr-evex-rtm \
076f82
+  rawmemchr-sse2 \
076f82
+  stpcpy-avx2 \
076f82
+  stpcpy-avx2-rtm \
076f82
+  stpcpy-evex \
076f82
+  stpcpy-sse2 \
076f82
+  stpcpy-sse2-unaligned \
076f82
+  stpcpy-ssse3 \
076f82
+  stpncpy-avx2 \
076f82
+  stpncpy-avx2-rtm \
076f82
+  stpncpy-c \
076f82
+  stpncpy-evex \
076f82
+  stpncpy-sse2-unaligned \
076f82
+  stpncpy-ssse3 \
076f82
+  strcasecmp_l-avx \
076f82
+  strcasecmp_l-sse2 \
076f82
+  strcasecmp_l-sse4_2 \
076f82
+  strcasecmp_l-ssse3 \
076f82
+  strcat-avx2 \
076f82
+  strcat-avx2-rtm \
076f82
+  strcat-evex \
076f82
+  strcat-sse2 \
076f82
+  strcat-sse2-unaligned \
076f82
+  strcat-ssse3 \
076f82
+  strchr-avx2 \
076f82
+  strchr-avx2-rtm \
076f82
+  strchr-evex \
076f82
+  strchr-sse2 \
076f82
+  strchr-sse2-no-bsf \
076f82
+  strchrnul-avx2 \
076f82
+  strchrnul-avx2-rtm \
076f82
+  strchrnul-evex \
076f82
+  strchrnul-sse2 \
076f82
+  strcmp-avx2 \
076f82
+  strcmp-avx2-rtm \
076f82
+  strcmp-evex \
076f82
+  strcmp-sse2 \
076f82
+  strcmp-sse2-unaligned \
076f82
+  strcmp-sse4_2 \
076f82
+  strcmp-ssse3 \
076f82
+  strcpy-avx2 \
076f82
+  strcpy-avx2-rtm \
076f82
+  strcpy-evex \
076f82
+  strcpy-sse2 \
076f82
+  strcpy-sse2-unaligned \
076f82
+  strcpy-ssse3 \
076f82
+  strcspn-c \
076f82
+  strcspn-sse2 \
076f82
+  strlen-avx2 \
076f82
+  strlen-avx2-rtm \
076f82
+  strlen-evex \
076f82
+  strlen-sse2 \
076f82
+  strncase_l-avx \
076f82
+  strncase_l-sse2 \
076f82
+  strncase_l-sse4_2 \
076f82
+  strncase_l-ssse3 \
076f82
+  strncat-avx2 \
076f82
+  strncat-avx2-rtm \
076f82
+  strncat-c \
076f82
+  strncat-evex \
076f82
+  strncat-sse2-unaligned \
076f82
+  strncat-ssse3 \
076f82
+  strncmp-avx2 \
076f82
+  strncmp-avx2-rtm \
076f82
+  strncmp-evex \
076f82
+  strncmp-sse2 \
076f82
+  strncmp-sse4_2 \
076f82
+  strncmp-ssse3 \
076f82
+  strncpy-avx2 \
076f82
+  strncpy-avx2-rtm \
076f82
+  strncpy-c \
076f82
+  strncpy-evex \
076f82
+  strncpy-sse2-unaligned \
076f82
+  strncpy-ssse3 \
076f82
+  strnlen-avx2 \
076f82
+  strnlen-avx2-rtm \
076f82
+  strnlen-evex \
076f82
+  strnlen-sse2 \
076f82
+  strpbrk-c \
076f82
+  strpbrk-sse2 \
076f82
+  strrchr-avx2 \
076f82
+  strrchr-avx2-rtm \
076f82
+  strrchr-evex \
076f82
+  strrchr-sse2 \
076f82
+  strspn-c \
076f82
+  strspn-sse2 \
076f82
+  strstr-sse2-unaligned \
076f82
+  varshift \
076f82
+# sysdep_routines
076f82
 CFLAGS-varshift.c += -msse4
076f82
 CFLAGS-strcspn-c.c += -msse4
076f82
 CFLAGS-strpbrk-c.c += -msse4
076f82
diff --git a/sysdeps/x86_64/multiarch/bzero.c b/sysdeps/x86_64/multiarch/bzero.c
076f82
new file mode 100644
076f82
index 0000000000000000..13e399a9a1fbdeb2
076f82
--- /dev/null
076f82
+++ b/sysdeps/x86_64/multiarch/bzero.c
076f82
@@ -0,0 +1,108 @@
076f82
+/* Multiple versions of bzero.
076f82
+   All versions must be listed in ifunc-impl-list.c.
076f82
+   Copyright (C) 2022 Free Software Foundation, Inc.
076f82
+   This file is part of the GNU C Library.
076f82
+
076f82
+   The GNU C Library is free software; you can redistribute it and/or
076f82
+   modify it under the terms of the GNU Lesser General Public
076f82
+   License as published by the Free Software Foundation; either
076f82
+   version 2.1 of the License, or (at your option) any later version.
076f82
+
076f82
+   The GNU C Library is distributed in the hope that it will be useful,
076f82
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
076f82
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
076f82
+   Lesser General Public License for more details.
076f82
+
076f82
+   You should have received a copy of the GNU Lesser General Public
076f82
+   License along with the GNU C Library; if not, see
076f82
+   <https://www.gnu.org/licenses/>.  */
076f82
+
076f82
+/* Define multiple versions only for the definition in libc.  */
076f82
+#if IS_IN (libc)
076f82
+# define __bzero __redirect___bzero
076f82
+# include <string.h>
076f82
+# undef __bzero
076f82
+
076f82
+/* OPTIMIZE1 definition required for bzero patch.  */
076f82
+# define OPTIMIZE1(name)	EVALUATOR1 (SYMBOL_NAME, name)
076f82
+# define SYMBOL_NAME __bzero
076f82
+# include <init-arch.h>
076f82
+
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (sse2_unaligned)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (sse2_unaligned_erms)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (avx2_unaligned) attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (avx2_unaligned_erms)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (avx2_unaligned_rtm)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (avx2_unaligned_erms_rtm)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (evex_unaligned)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (evex_unaligned_erms)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (avx512_unaligned)
076f82
+  attribute_hidden;
076f82
+extern __typeof (REDIRECT_NAME) OPTIMIZE1 (avx512_unaligned_erms)
076f82
+  attribute_hidden;
076f82
+
076f82
+static inline void *
076f82
+IFUNC_SELECTOR (void)
076f82
+{
076f82
+  const struct cpu_features* cpu_features = __get_cpu_features ();
076f82
+
076f82
+  if (CPU_FEATURE_USABLE_P (cpu_features, AVX512F)
076f82
+      && !CPU_FEATURES_ARCH_P (cpu_features, Prefer_No_AVX512))
076f82
+    {
076f82
+      if (CPU_FEATURE_USABLE_P (cpu_features, AVX512VL)
076f82
+          && CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
076f82
+          && CPU_FEATURE_USABLE_P (cpu_features, BMI2))
076f82
+	{
076f82
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
076f82
+	    return OPTIMIZE1 (avx512_unaligned_erms);
076f82
+
076f82
+	  return OPTIMIZE1 (avx512_unaligned);
076f82
+	}
076f82
+    }
076f82
+
076f82
+  if (CPU_FEATURE_USABLE_P (cpu_features, AVX2))
076f82
+    {
076f82
+      if (CPU_FEATURE_USABLE_P (cpu_features, AVX512VL)
076f82
+          && CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
076f82
+          && CPU_FEATURE_USABLE_P (cpu_features, BMI2))
076f82
+	{
076f82
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
076f82
+	    return OPTIMIZE1 (evex_unaligned_erms);
076f82
+
076f82
+	  return OPTIMIZE1 (evex_unaligned);
076f82
+	}
076f82
+
076f82
+      if (CPU_FEATURE_USABLE_P (cpu_features, RTM))
076f82
+	{
076f82
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
076f82
+	    return OPTIMIZE1 (avx2_unaligned_erms_rtm);
076f82
+
076f82
+	  return OPTIMIZE1 (avx2_unaligned_rtm);
076f82
+	}
076f82
+
076f82
+      if (!CPU_FEATURES_ARCH_P (cpu_features, Prefer_No_VZEROUPPER))
076f82
+	{
076f82
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
076f82
+	    return OPTIMIZE1 (avx2_unaligned_erms);
076f82
+
076f82
+	  return OPTIMIZE1 (avx2_unaligned);
076f82
+	}
076f82
+    }
076f82
+
076f82
+  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
076f82
+    return OPTIMIZE1 (sse2_unaligned_erms);
076f82
+
076f82
+  return OPTIMIZE1 (sse2_unaligned);
076f82
+}
076f82
+
076f82
+libc_ifunc_redirected (__redirect___bzero, __bzero, IFUNC_SELECTOR ());
076f82
+
076f82
+weak_alias (__bzero, bzero)
076f82
+#endif
076f82
diff --git a/sysdeps/x86_64/multiarch/ifunc-impl-list.c b/sysdeps/x86_64/multiarch/ifunc-impl-list.c
076f82
index 39ab10613bb0ffea..4992d7bd3206a7c0 100644
076f82
--- a/sysdeps/x86_64/multiarch/ifunc-impl-list.c
076f82
+++ b/sysdeps/x86_64/multiarch/ifunc-impl-list.c
076f82
@@ -282,6 +282,48 @@ __libc_ifunc_impl_list (const char *name, struct libc_ifunc_impl *array,
076f82
 			      __memset_avx512_no_vzeroupper)
076f82
 	     )
076f82
 
076f82
+  /* Support sysdeps/x86_64/multiarch/bzero.c.  */
076f82
+  IFUNC_IMPL (i, name, bzero,
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero, 1,
076f82
+			      __bzero_sse2_unaligned)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero, 1,
076f82
+			      __bzero_sse2_unaligned_erms)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      CPU_FEATURE_USABLE (AVX2),
076f82
+			      __bzero_avx2_unaligned)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      CPU_FEATURE_USABLE (AVX2),
076f82
+			      __bzero_avx2_unaligned_erms)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      (CPU_FEATURE_USABLE (AVX2)
076f82
+			       && CPU_FEATURE_USABLE (RTM)),
076f82
+			      __bzero_avx2_unaligned_rtm)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      (CPU_FEATURE_USABLE (AVX2)
076f82
+			       && CPU_FEATURE_USABLE (RTM)),
076f82
+			      __bzero_avx2_unaligned_erms_rtm)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      (CPU_FEATURE_USABLE (AVX512VL)
076f82
+			       && CPU_FEATURE_USABLE (AVX512BW)
076f82
+			       && CPU_FEATURE_USABLE (BMI2)),
076f82
+			      __bzero_evex_unaligned)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      (CPU_FEATURE_USABLE (AVX512VL)
076f82
+			       && CPU_FEATURE_USABLE (AVX512BW)
076f82
+			       && CPU_FEATURE_USABLE (BMI2)),
076f82
+			      __bzero_evex_unaligned_erms)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      (CPU_FEATURE_USABLE (AVX512VL)
076f82
+			       && CPU_FEATURE_USABLE (AVX512BW)
076f82
+			       && CPU_FEATURE_USABLE (BMI2)),
076f82
+			      __bzero_avx512_unaligned_erms)
076f82
+	      IFUNC_IMPL_ADD (array, i, bzero,
076f82
+			      (CPU_FEATURE_USABLE (AVX512VL)
076f82
+			       && CPU_FEATURE_USABLE (AVX512BW)
076f82
+			       && CPU_FEATURE_USABLE (BMI2)),
076f82
+			      __bzero_avx512_unaligned)
076f82
+	     )
076f82
+
076f82
   /* Support sysdeps/x86_64/multiarch/rawmemchr.c.  */
076f82
   IFUNC_IMPL (i, name, rawmemchr,
076f82
 	      IFUNC_IMPL_ADD (array, i, rawmemchr,
076f82
diff --git a/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms-rtm.S b/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms-rtm.S
076f82
index 8ac3e479bba488be..5a5ee6f67299400b 100644
076f82
--- a/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms-rtm.S
076f82
+++ b/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms-rtm.S
076f82
@@ -5,6 +5,7 @@
076f82
 
076f82
 #define SECTION(p) p##.avx.rtm
076f82
 #define MEMSET_SYMBOL(p,s)	p##_avx2_##s##_rtm
076f82
+#define BZERO_SYMBOL(p,s)	p##_avx2_##s##_rtm
076f82
 #define WMEMSET_SYMBOL(p,s)	p##_avx2_##s##_rtm
076f82
 
076f82
 #include "memset-avx2-unaligned-erms.S"
076f82
diff --git a/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms.S b/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms.S
076f82
index c0bf2875d03d51ab..a093a2831f3dfa0d 100644
076f82
--- a/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms.S
076f82
+++ b/sysdeps/x86_64/multiarch/memset-avx2-unaligned-erms.S
076f82
@@ -14,6 +14,9 @@
076f82
   vmovd d, %xmm0; \
076f82
   movq r, %rax;
076f82
 
076f82
+# define BZERO_ZERO_VEC0() \
076f82
+  vpxor %xmm0, %xmm0, %xmm0
076f82
+
076f82
 # define WMEMSET_SET_VEC0_AND_SET_RETURN(d, r) \
076f82
   MEMSET_SET_VEC0_AND_SET_RETURN(d, r)
076f82
 
076f82
@@ -29,6 +32,9 @@
076f82
 # ifndef MEMSET_SYMBOL
076f82
 #  define MEMSET_SYMBOL(p,s)	p##_avx2_##s
076f82
 # endif
076f82
+# ifndef BZERO_SYMBOL
076f82
+#  define BZERO_SYMBOL(p,s)	p##_avx2_##s
076f82
+# endif
076f82
 # ifndef WMEMSET_SYMBOL
076f82
 #  define WMEMSET_SYMBOL(p,s)	p##_avx2_##s
076f82
 # endif
076f82
diff --git a/sysdeps/x86_64/multiarch/memset-avx512-unaligned-erms.S b/sysdeps/x86_64/multiarch/memset-avx512-unaligned-erms.S
076f82
index 5241216a77bf72b7..727c92133a15900f 100644
076f82
--- a/sysdeps/x86_64/multiarch/memset-avx512-unaligned-erms.S
076f82
+++ b/sysdeps/x86_64/multiarch/memset-avx512-unaligned-erms.S
076f82
@@ -19,6 +19,9 @@
076f82
   vpbroadcastb d, %VEC0; \
076f82
   movq r, %rax
076f82
 
076f82
+# define BZERO_ZERO_VEC0() \
076f82
+  vpxorq %XMM0, %XMM0, %XMM0
076f82
+
076f82
 # define WMEMSET_SET_VEC0_AND_SET_RETURN(d, r) \
076f82
   vpbroadcastd d, %VEC0; \
076f82
   movq r, %rax
076f82
diff --git a/sysdeps/x86_64/multiarch/memset-evex-unaligned-erms.S b/sysdeps/x86_64/multiarch/memset-evex-unaligned-erms.S
076f82
index 637002150659123c..5d8fa78f05476b10 100644
076f82
--- a/sysdeps/x86_64/multiarch/memset-evex-unaligned-erms.S
076f82
+++ b/sysdeps/x86_64/multiarch/memset-evex-unaligned-erms.S
076f82
@@ -19,6 +19,9 @@
076f82
   vpbroadcastb d, %VEC0; \
076f82
   movq r, %rax
076f82
 
076f82
+# define BZERO_ZERO_VEC0() \
076f82
+  vpxorq %XMM0, %XMM0, %XMM0
076f82
+
076f82
 # define WMEMSET_SET_VEC0_AND_SET_RETURN(d, r) \
076f82
   vpbroadcastd d, %VEC0; \
076f82
   movq r, %rax
076f82
diff --git a/sysdeps/x86_64/multiarch/memset-sse2-unaligned-erms.S b/sysdeps/x86_64/multiarch/memset-sse2-unaligned-erms.S
076f82
index e4e95fc19fe48d2d..bac74ac37fd3c144 100644
076f82
--- a/sysdeps/x86_64/multiarch/memset-sse2-unaligned-erms.S
076f82
+++ b/sysdeps/x86_64/multiarch/memset-sse2-unaligned-erms.S
076f82
@@ -22,6 +22,7 @@
076f82
 
076f82
 #if IS_IN (libc)
076f82
 # define MEMSET_SYMBOL(p,s)	p##_sse2_##s
076f82
+# define BZERO_SYMBOL(p,s)	MEMSET_SYMBOL (p, s)
076f82
 # define WMEMSET_SYMBOL(p,s)	p##_sse2_##s
076f82
 
076f82
 # ifdef SHARED
076f82
diff --git a/sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S
076f82
index c8db87dcbf69f0d8..39a096a594ccb5b6 100644
076f82
--- a/sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S
076f82
+++ b/sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S
076f82
@@ -26,6 +26,10 @@
076f82
 
076f82
 #include <sysdep.h>
076f82
 
076f82
+#ifndef BZERO_SYMBOL
076f82
+# define BZERO_SYMBOL(p,s)		MEMSET_SYMBOL (p, s)
076f82
+#endif
076f82
+
076f82
 #ifndef MEMSET_CHK_SYMBOL
076f82
 # define MEMSET_CHK_SYMBOL(p,s)		MEMSET_SYMBOL(p, s)
076f82
 #endif
076f82
@@ -87,6 +91,18 @@
076f82
 # define XMM_SMALL	0
076f82
 #endif
076f82
 
076f82
+#ifdef USE_LESS_VEC_MASK_STORE
076f82
+# define SET_REG64	rcx
076f82
+# define SET_REG32	ecx
076f82
+# define SET_REG16	cx
076f82
+# define SET_REG8	cl
076f82
+#else
076f82
+# define SET_REG64	rsi
076f82
+# define SET_REG32	esi
076f82
+# define SET_REG16	si
076f82
+# define SET_REG8	sil
076f82
+#endif
076f82
+
076f82
 #define PAGE_SIZE 4096
076f82
 
076f82
 /* Macro to calculate size of small memset block for aligning
076f82
@@ -96,18 +112,6 @@
076f82
 
076f82
 #ifndef SECTION
076f82
 # error SECTION is not defined!
076f82
-#endif
076f82
-
076f82
-	.section SECTION(.text),"ax",@progbits
076f82
-#if VEC_SIZE == 16 && IS_IN (libc)
076f82
-ENTRY (__bzero)
076f82
-	mov	%RDI_LP, %RAX_LP /* Set return value.  */
076f82
-	mov	%RSI_LP, %RDX_LP /* Set n.  */
076f82
-	xorl	%esi, %esi
076f82
-	pxor	%XMM0, %XMM0
076f82
-	jmp	L(entry_from_bzero)
076f82
-END (__bzero)
076f82
-weak_alias (__bzero, bzero)
076f82
 #endif
076f82
 
076f82
 #if IS_IN (libc)
076f82
@@ -123,12 +127,37 @@ ENTRY (WMEMSET_SYMBOL (__wmemset, unaligned))
076f82
 	WMEMSET_SET_VEC0_AND_SET_RETURN (%esi, %rdi)
076f82
 	WMEMSET_VDUP_TO_VEC0_LOW()
076f82
 	cmpq	$VEC_SIZE, %rdx
076f82
-	jb	L(less_vec_no_vdup)
076f82
+	jb	L(less_vec_from_wmemset)
076f82
 	WMEMSET_VDUP_TO_VEC0_HIGH()
076f82
 	jmp	L(entry_from_wmemset)
076f82
 END (WMEMSET_SYMBOL (__wmemset, unaligned))
076f82
 #endif
076f82
 
076f82
+ENTRY (BZERO_SYMBOL(__bzero, unaligned))
076f82
+#if VEC_SIZE > 16
076f82
+	BZERO_ZERO_VEC0 ()
076f82
+#endif
076f82
+	mov	%RDI_LP, %RAX_LP
076f82
+	mov	%RSI_LP, %RDX_LP
076f82
+#ifndef USE_LESS_VEC_MASK_STORE
076f82
+	xorl	%esi, %esi
076f82
+#endif
076f82
+	cmp	$VEC_SIZE, %RDX_LP
076f82
+	jb	L(less_vec_no_vdup)
076f82
+#ifdef USE_LESS_VEC_MASK_STORE
076f82
+	xorl	%esi, %esi
076f82
+#endif
076f82
+#if VEC_SIZE <= 16
076f82
+	BZERO_ZERO_VEC0 ()
076f82
+#endif
076f82
+	cmp	$(VEC_SIZE * 2), %RDX_LP
076f82
+	ja	L(more_2x_vec)
076f82
+	/* From VEC and to 2 * VEC.  No branch when size == VEC_SIZE.  */
076f82
+	VMOVU	%VEC(0), (%rdi)
076f82
+	VMOVU	%VEC(0), (VEC_SIZE * -1)(%rdi, %rdx)
076f82
+	VZEROUPPER_RETURN
076f82
+END (BZERO_SYMBOL(__bzero, unaligned))
076f82
+
076f82
 #if defined SHARED && IS_IN (libc)
076f82
 ENTRY_CHK (MEMSET_CHK_SYMBOL (__memset_chk, unaligned))
076f82
 	cmp	%RDX_LP, %RCX_LP
076f82
@@ -142,7 +171,6 @@ ENTRY (MEMSET_SYMBOL (__memset, unaligned))
076f82
 	/* Clear the upper 32 bits.  */
076f82
 	mov	%edx, %edx
076f82
 # endif
076f82
-L(entry_from_bzero):
076f82
 	cmpq	$VEC_SIZE, %rdx
076f82
 	jb	L(less_vec)
076f82
 	MEMSET_VDUP_TO_VEC0_HIGH()
076f82
@@ -187,6 +215,31 @@ END (__memset_erms)
076f82
 END (MEMSET_SYMBOL (__memset, erms))
076f82
 # endif
076f82
 
076f82
+ENTRY_P2ALIGN (BZERO_SYMBOL(__bzero, unaligned_erms), 6)
076f82
+# if VEC_SIZE > 16
076f82
+	BZERO_ZERO_VEC0 ()
076f82
+# endif
076f82
+	mov	%RDI_LP, %RAX_LP
076f82
+	mov	%RSI_LP, %RDX_LP
076f82
+# ifndef USE_LESS_VEC_MASK_STORE
076f82
+	xorl	%esi, %esi
076f82
+# endif
076f82
+	cmp	$VEC_SIZE, %RDX_LP
076f82
+	jb	L(less_vec_no_vdup)
076f82
+# ifdef USE_LESS_VEC_MASK_STORE
076f82
+	xorl	%esi, %esi
076f82
+# endif
076f82
+# if VEC_SIZE <= 16
076f82
+	BZERO_ZERO_VEC0 ()
076f82
+# endif
076f82
+	cmp	$(VEC_SIZE * 2), %RDX_LP
076f82
+	ja	L(stosb_more_2x_vec)
076f82
+	/* From VEC and to 2 * VEC.  No branch when size == VEC_SIZE.  */
076f82
+	VMOVU	%VEC(0), (%rdi)
076f82
+	VMOVU	%VEC(0), (VEC_SIZE * -1)(%rdi, %rdx)
076f82
+	VZEROUPPER_RETURN
076f82
+END (BZERO_SYMBOL(__bzero, unaligned_erms))
076f82
+
076f82
 # if defined SHARED && IS_IN (libc)
076f82
 ENTRY_CHK (MEMSET_CHK_SYMBOL (__memset_chk, unaligned_erms))
076f82
 	cmp	%RDX_LP, %RCX_LP
076f82
@@ -229,6 +282,7 @@ L(last_2x_vec):
076f82
 	.p2align 4,, 10
076f82
 L(less_vec):
076f82
 L(less_vec_no_vdup):
076f82
+L(less_vec_from_wmemset):
076f82
 	/* Less than 1 VEC.  */
076f82
 # if VEC_SIZE != 16 && VEC_SIZE != 32 && VEC_SIZE != 64
076f82
 #  error Unsupported VEC_SIZE!
076f82
@@ -374,8 +428,11 @@ L(less_vec):
076f82
 	/* Broadcast esi to partial register (i.e VEC_SIZE == 32 broadcast to
076f82
 	   xmm). This is only does anything for AVX2.  */
076f82
 	MEMSET_VDUP_TO_VEC0_LOW ()
076f82
+L(less_vec_from_wmemset):
076f82
+#if VEC_SIZE > 16
076f82
 L(less_vec_no_vdup):
076f82
 #endif
076f82
+#endif
076f82
 L(cross_page):
076f82
 #if VEC_SIZE > 32
076f82
 	cmpl	$32, %edx
076f82
@@ -386,7 +443,10 @@ L(cross_page):
076f82
 	jge	L(between_16_31)
076f82
 #endif
076f82
 #ifndef USE_XMM_LESS_VEC
076f82
-	MOVQ	%XMM0, %rcx
076f82
+	MOVQ	%XMM0, %SET_REG64
076f82
+#endif
076f82
+#if VEC_SIZE <= 16
076f82
+L(less_vec_no_vdup):
076f82
 #endif
076f82
 	cmpl	$8, %edx
076f82
 	jge	L(between_8_15)
076f82
@@ -395,7 +455,7 @@ L(cross_page):
076f82
 	cmpl	$1, %edx
076f82
 	jg	L(between_2_3)
076f82
 	jl	L(between_0_0)
076f82
-	movb	%sil, (%LESS_VEC_REG)
076f82
+	movb	%SET_REG8, (%LESS_VEC_REG)
076f82
 L(between_0_0):
076f82
 	ret
076f82
 
076f82
@@ -428,8 +488,8 @@ L(between_8_15):
076f82
 	MOVQ	%XMM0, (%rdi)
076f82
 	MOVQ	%XMM0, -8(%rdi, %rdx)
076f82
 #else
076f82
-	movq	%rcx, (%LESS_VEC_REG)
076f82
-	movq	%rcx, -8(%LESS_VEC_REG, %rdx)
076f82
+	movq	%SET_REG64, (%LESS_VEC_REG)
076f82
+	movq	%SET_REG64, -8(%LESS_VEC_REG, %rdx)
076f82
 #endif
076f82
 	ret
076f82
 
076f82
@@ -442,8 +502,8 @@ L(between_4_7):
076f82
 	MOVD	%XMM0, (%rdi)
076f82
 	MOVD	%XMM0, -4(%rdi, %rdx)
076f82
 #else
076f82
-	movl	%ecx, (%LESS_VEC_REG)
076f82
-	movl	%ecx, -4(%LESS_VEC_REG, %rdx)
076f82
+	movl	%SET_REG32, (%LESS_VEC_REG)
076f82
+	movl	%SET_REG32, -4(%LESS_VEC_REG, %rdx)
076f82
 #endif
076f82
 	ret
076f82
 
076f82
@@ -452,12 +512,12 @@ L(between_4_7):
076f82
 L(between_2_3):
076f82
 	/* From 2 to 3.  No branch when size == 2.  */
076f82
 #ifdef USE_XMM_LESS_VEC
076f82
-	movb	%sil, (%rdi)
076f82
-	movb	%sil, 1(%rdi)
076f82
-	movb	%sil, -1(%rdi, %rdx)
076f82
+	movb	%SET_REG8, (%rdi)
076f82
+	movb	%SET_REG8, 1(%rdi)
076f82
+	movb	%SET_REG8, -1(%rdi, %rdx)
076f82
 #else
076f82
-	movw	%cx, (%LESS_VEC_REG)
076f82
-	movb	%sil, -1(%LESS_VEC_REG, %rdx)
076f82
+	movw	%SET_REG16, (%LESS_VEC_REG)
076f82
+	movb	%SET_REG8, -1(%LESS_VEC_REG, %rdx)
076f82
 #endif
076f82
 	ret
076f82
 END (MEMSET_SYMBOL (__memset, unaligned_erms))