7486fd
diff --git a/include/private/gcconfig.h b/include/private/gcconfig.h
7486fd
index 767859c..ec92882 100644
7486fd
--- a/include/private/gcconfig.h
7486fd
+++ b/include/private/gcconfig.h
7486fd
@@ -70,6 +70,13 @@
7486fd
 #    define I386
7486fd
 #    define mach_type_known
7486fd
 # endif
7486fd
+# if defined(__aarch64__)
7486fd
+#    define AARCH64
7486fd
+#    if !defined(LINUX)
7486fd
+#      define NOSYS
7486fd
+#      define mach_type_known
7486fd
+#    endif
7486fd
+# endif
7486fd
 # if defined(__arm) || defined(__arm__) || defined(__thumb__)
7486fd
 #    define ARM32
7486fd
 #    if !defined(LINUX) && !defined(NETBSD) && !defined(OPENBSD) \
7486fd
@@ -250,6 +257,10 @@
7486fd
 #    define IA64
7486fd
 #    define mach_type_known
7486fd
 # endif
7486fd
+# if defined(LINUX) && defined(__aarch64__)
7486fd
+#    define AARCH64
7486fd
+#    define mach_type_known
7486fd
+# endif
7486fd
 # if defined(LINUX) && (defined(__arm) || defined(__arm__))
7486fd
 #    define ARM32
7486fd
 #    define mach_type_known
7486fd
@@ -537,6 +548,7 @@
7486fd
                     /*                  running Amdahl UTS4             */
7486fd
                     /*             S390       ==> 390-like machine      */
7486fd
                     /*                  running LINUX                   */
7486fd
+                    /*             AARCH64    ==> ARM AArch64           */
7486fd
                     /*             ARM32      ==> Intel StrongARM       */
7486fd
                     /*             IA64       ==> Intel IPF             */
7486fd
                     /*                            (e.g. Itanium)        */
7486fd
@@ -1899,6 +1911,31 @@
7486fd
 #   endif
7486fd
 # endif
7486fd
 
7486fd
+# ifdef AARCH64
7486fd
+#   define CPP_WORDSZ 64
7486fd
+#   define MACH_TYPE "AARCH64"
7486fd
+#   define ALIGNMENT 8
7486fd
+#   ifndef HBLKSIZE
7486fd
+#     define HBLKSIZE 4096
7486fd
+#   endif
7486fd
+#   ifdef LINUX
7486fd
+#     define OS_TYPE "LINUX"
7486fd
+#     define LINUX_STACKBOTTOM
7486fd
+#     define DYNAMIC_LOADING
7486fd
+      extern int __data_start[];
7486fd
+#     define DATASTART ((ptr_t)__data_start)
7486fd
+      extern char _end[];
7486fd
+#     define DATAEND ((ptr_t)(&_end))
7486fd
+#   endif
7486fd
+#   ifdef NOSYS
7486fd
+      /* __data_start is usually defined in the target linker script.   */
7486fd
+      extern int __data_start[];
7486fd
+#     define DATASTART ((ptr_t)__data_start)
7486fd
+      extern void *__stack_base__;
7486fd
+#     define STACKBOTTOM ((ptr_t)__stack_base__)
7486fd
+#   endif
7486fd
+# endif
7486fd
+
7486fd
 # ifdef ARM32
7486fd
 #   define CPP_WORDSZ 32
7486fd
 #   define MACH_TYPE "ARM32"
7486fd
diff --git a/libatomic_ops/src/atomic_ops.h b/libatomic_ops/src/atomic_ops.h
7486fd
index db177d5..d91da53 100644
7486fd
--- a/libatomic_ops/src/atomic_ops.h
7486fd
+++ b/libatomic_ops/src/atomic_ops.h
7486fd
@@ -244,6 +244,10 @@
7486fd
      || defined(__powerpc64__) || defined(__ppc64__)
7486fd
 #   include "atomic_ops/sysdeps/gcc/powerpc.h"
7486fd
 # endif /* __powerpc__ */
7486fd
+# if defined(__aarch64__)
7486fd
+#   include "atomic_ops/sysdeps/gcc/aarch64.h"
7486fd
+#   define AO_CAN_EMUL_CAS
7486fd
+# endif /* __aarch64__ */
7486fd
 # if defined(__arm__) && !defined(AO_USE_PTHREAD_DEFS)
7486fd
 #   include "atomic_ops/sysdeps/gcc/arm.h"
7486fd
 #   define AO_CAN_EMUL_CAS
7486fd
diff --git a/libatomic_ops/src/atomic_ops/sysdeps/Makefile.am b/libatomic_ops/src/atomic_ops/sysdeps/Makefile.am
7486fd
index d8b24dc..b73a20c 100644
7486fd
--- a/libatomic_ops/src/atomic_ops/sysdeps/Makefile.am
7486fd
+++ b/libatomic_ops/src/atomic_ops/sysdeps/Makefile.am
7486fd
@@ -30,6 +30,7 @@ nobase_sysdep_HEADERS= generic_pthread.h \
7486fd
 	  gcc/hexagon.h gcc/hppa.h gcc/ia64.h gcc/m68k.h \
7486fd
 	  gcc/mips.h gcc/powerpc.h gcc/s390.h \
7486fd
 	  gcc/sh.h gcc/sparc.h gcc/x86.h gcc/x86_64.h \
7486fd
+	  gcc/aarch64.h \
7486fd
 	\
7486fd
 	  hpc/hppa.h hpc/ia64.h \
7486fd
 	\
7486fd
diff --git a/libatomic_ops/src/atomic_ops/sysdeps/gcc/aarch64.h b/libatomic_ops/src/atomic_ops/sysdeps/gcc/aarch64.h
7486fd
new file mode 100644
7486fd
index 0000000..94f1f14
7486fd
--- /dev/null
7486fd
+++ b/libatomic_ops/src/atomic_ops/sysdeps/gcc/aarch64.h
7486fd
@@ -0,0 +1,353 @@
7486fd
+/*
7486fd
+ * Copyright (c) 1991-1994 by Xerox Corporation.  All rights reserved.
7486fd
+ * Copyright (c) 1996-1999 by Silicon Graphics.  All rights reserved.
7486fd
+ * Copyright (c) 1999-2003 by Hewlett-Packard Company. All rights reserved.
7486fd
+ *
7486fd
+ *
7486fd
+ * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED
7486fd
+ * OR IMPLIED.  ANY USE IS AT YOUR OWN RISK.
7486fd
+ *
7486fd
+ * Permission is hereby granted to use or copy this program
7486fd
+ * for any purpose,  provided the above notices are retained on all copies.
7486fd
+ * Permission to modify the code and to distribute modified code is granted,
7486fd
+ * provided the above notices are retained, and a notice that the code was
7486fd
+ * modified is included with the above copyright notice.
7486fd
+ *
7486fd
+ */
7486fd
+
7486fd
+#include "../read_ordered.h"
7486fd
+
7486fd
+#include "../test_and_set_t_is_ao_t.h"
7486fd
+
7486fd
+#include "../standard_ao_double_t.h"
7486fd
+
7486fd
+#ifndef AO_UNIPROCESSOR
7486fd
+  AO_INLINE void
7486fd
+  AO_nop_write(void)
7486fd
+  {
7486fd
+    __asm__ __volatile__("dmb st" : : : "memory");
7486fd
+  }
7486fd
+# define AO_HAVE_nop_write
7486fd
+#endif
7486fd
+
7486fd
+#ifndef AO_EXPECT_FALSE
7486fd
+#if __GNUC__ >= 3 && !defined(LINT2)
7486fd
+# define AO_EXPECT_FALSE(expr) __builtin_expect(expr, 0)
7486fd
+  /* Equivalent to (expr) but predict that usually (expr) == 0. */
7486fd
+#else
7486fd
+# define AO_EXPECT_FALSE(expr) (expr)
7486fd
+#endif /* !__GNUC__ */
7486fd
+#endif
7486fd
+
7486fd
+/* TODO: Adjust version check on fixing double-wide AO support in GCC. */
7486fd
+#if __GNUC__ == 4
7486fd
+
7486fd
+  AO_INLINE AO_double_t
7486fd
+  AO_double_load(const volatile AO_double_t *addr)
7486fd
+  {
7486fd
+    AO_double_t result;
7486fd
+    int status;
7486fd
+
7486fd
+    /* Note that STXP cannot be discarded because LD[A]XP is not        */
7486fd
+    /* single-copy atomic (unlike LDREXD for 32-bit ARM).               */
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_load\n"
7486fd
+      "       ldxp  %0, %1, %3\n"
7486fd
+      "       stxp %w2, %0, %1, %3"
7486fd
+      : "=&r" (result.AO_val1), "=&r" (result.AO_val2), "=&r" (status)
7486fd
+      : "Q" (*addr));
7486fd
+    } while (AO_EXPECT_FALSE(status));
7486fd
+    return result;
7486fd
+  }
7486fd
+# define AO_HAVE_double_load
7486fd
+
7486fd
+  AO_INLINE AO_double_t
7486fd
+  AO_double_load_acquire(const volatile AO_double_t *addr)
7486fd
+  {
7486fd
+    AO_double_t result;
7486fd
+    int status;
7486fd
+
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_load_acquire\n"
7486fd
+      "       ldaxp  %0, %1, %3\n"
7486fd
+      "       stxp %w2, %0, %1, %3"
7486fd
+      : "=&r" (result.AO_val1), "=&r" (result.AO_val2), "=&r" (status)
7486fd
+      : "Q" (*addr));
7486fd
+    } while (AO_EXPECT_FALSE(status));
7486fd
+    return result;
7486fd
+  }
7486fd
+# define AO_HAVE_double_load_acquire
7486fd
+
7486fd
+  AO_INLINE void
7486fd
+  AO_double_store(volatile AO_double_t *addr, AO_double_t value)
7486fd
+  {
7486fd
+    AO_double_t old_val;
7486fd
+    int status;
7486fd
+
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_store\n"
7486fd
+      "       ldxp  %0, %1, %3\n"
7486fd
+      "       stxp %w2, %4, %5, %3"
7486fd
+      : "=&r" (old_val.AO_val1), "=&r" (old_val.AO_val2), "=&r" (status),
7486fd
+        "=Q" (*addr)
7486fd
+      : "r" (value.AO_val1), "r" (value.AO_val2));
7486fd
+      /* Compared to the arm.h implementation, the 'cc' (flags) are not */
7486fd
+      /* clobbered because A64 has no concept of conditional execution. */
7486fd
+    } while (AO_EXPECT_FALSE(status));
7486fd
+  }
7486fd
+# define AO_HAVE_double_store
7486fd
+
7486fd
+  AO_INLINE void
7486fd
+  AO_double_store_release(volatile AO_double_t *addr, AO_double_t value)
7486fd
+  {
7486fd
+    AO_double_t old_val;
7486fd
+    int status;
7486fd
+
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_store_release\n"
7486fd
+      "       ldxp  %0, %1, %3\n"
7486fd
+      "       stlxp %w2, %4, %5, %3"
7486fd
+      : "=&r" (old_val.AO_val1), "=&r" (old_val.AO_val2), "=&r" (status),
7486fd
+        "=Q" (*addr)
7486fd
+      : "r" (value.AO_val1), "r" (value.AO_val2));
7486fd
+    } while (AO_EXPECT_FALSE(status));
7486fd
+  }
7486fd
+# define AO_HAVE_double_store_release
7486fd
+
7486fd
+  AO_INLINE int
7486fd
+  AO_double_compare_and_swap(volatile AO_double_t *addr,
7486fd
+                             AO_double_t old_val, AO_double_t new_val)
7486fd
+  {
7486fd
+    AO_double_t tmp;
7486fd
+    int result = 1;
7486fd
+
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_compare_and_swap\n"
7486fd
+        "       ldxp  %0, %1, %2\n"
7486fd
+        : "=&r" (tmp.AO_val1), "=&r" (tmp.AO_val2)
7486fd
+        : "Q" (*addr));
7486fd
+      if (tmp.AO_val1 != old_val.AO_val1 || tmp.AO_val2 != old_val.AO_val2)
7486fd
+        break;
7486fd
+      __asm__ __volatile__(
7486fd
+        "       stxp %w0, %2, %3, %1\n"
7486fd
+        : "=&r" (result), "=Q" (*addr)
7486fd
+        : "r" (new_val.AO_val1), "r" (new_val.AO_val2));
7486fd
+    } while (AO_EXPECT_FALSE(result));
7486fd
+    return !result;
7486fd
+  }
7486fd
+# define AO_HAVE_double_compare_and_swap
7486fd
+
7486fd
+  AO_INLINE int
7486fd
+  AO_double_compare_and_swap_acquire(volatile AO_double_t *addr,
7486fd
+                                     AO_double_t old_val, AO_double_t new_val)
7486fd
+  {
7486fd
+    AO_double_t tmp;
7486fd
+    int result = 1;
7486fd
+
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_compare_and_swap_acquire\n"
7486fd
+        "       ldaxp  %0, %1, %2\n"
7486fd
+        : "=&r" (tmp.AO_val1), "=&r" (tmp.AO_val2)
7486fd
+        : "Q" (*addr));
7486fd
+      if (tmp.AO_val1 != old_val.AO_val1 || tmp.AO_val2 != old_val.AO_val2)
7486fd
+        break;
7486fd
+      __asm__ __volatile__(
7486fd
+        "       stxp %w0, %2, %3, %1\n"
7486fd
+        : "=&r" (result), "=Q" (*addr)
7486fd
+        : "r" (new_val.AO_val1), "r" (new_val.AO_val2));
7486fd
+    } while (AO_EXPECT_FALSE(result));
7486fd
+    return !result;
7486fd
+  }
7486fd
+# define AO_HAVE_double_compare_and_swap_acquire
7486fd
+
7486fd
+  AO_INLINE int
7486fd
+  AO_double_compare_and_swap_release(volatile AO_double_t *addr,
7486fd
+                                     AO_double_t old_val, AO_double_t new_val)
7486fd
+  {
7486fd
+    AO_double_t tmp;
7486fd
+    int result = 1;
7486fd
+
7486fd
+    do {
7486fd
+      __asm__ __volatile__("//AO_double_compare_and_swap_release\n"
7486fd
+        "       ldxp  %0, %1, %2\n"
7486fd
+        : "=&r" (tmp.AO_val1), "=&r" (tmp.AO_val2)
7486fd
+        : "Q" (*addr));
7486fd
+      if (tmp.AO_val1 != old_val.AO_val1 || tmp.AO_val2 != old_val.AO_val2)
7486fd
+        break;
7486fd
+      __asm__ __volatile__(
7486fd
+        "       stlxp %w0, %2, %3, %1\n"
7486fd
+        : "=&r" (result), "=Q" (*addr)
7486fd
+        : "r" (new_val.AO_val1), "r" (new_val.AO_val2));
7486fd
+    } while (AO_EXPECT_FALSE(result));
7486fd
+    return !result;
7486fd
+  }
7486fd
+# define AO_HAVE_double_compare_and_swap_release
7486fd
+#endif
7486fd
+
7486fd
+AO_INLINE void
7486fd
+AO_nop_full(void)
7486fd
+{
7486fd
+# ifndef AO_UNIPROCESSOR
7486fd
+__sync_synchronize ();
7486fd
+# endif
7486fd
+}
7486fd
+#define AO_HAVE_nop_full
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_load(const volatile AO_t *addr)
7486fd
+{
7486fd
+  return  (AO_t)__atomic_load_n (addr, __ATOMIC_RELAXED);
7486fd
+}
7486fd
+#define AO_HAVE_load
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_load_acquire(const volatile AO_t *addr)
7486fd
+{
7486fd
+  return (AO_t)__atomic_load_n (addr, __ATOMIC_ACQUIRE);
7486fd
+}
7486fd
+#define AO_HAVE_load_acquire
7486fd
+
7486fd
+AO_INLINE void
7486fd
+ AO_store(volatile AO_t *addr, AO_t value)
7486fd
+{
7486fd
+  __atomic_store_n(addr, value, __ATOMIC_RELAXED);
7486fd
+}
7486fd
+#define AO_HAVE_store
7486fd
+
7486fd
+AO_INLINE void
7486fd
+ AO_store_release(volatile AO_t *addr, AO_t value)
7486fd
+{
7486fd
+  __atomic_store_n(addr, value, __ATOMIC_RELEASE);
7486fd
+}
7486fd
+#define AO_HAVE_store_release
7486fd
+
7486fd
+AO_INLINE AO_TS_VAL_t
7486fd
+AO_test_and_set(volatile AO_TS_t *addr)
7486fd
+{
7486fd
+  return (AO_TS_VAL_t)__atomic_test_and_set(addr, __ATOMIC_RELAXED);
7486fd
+}
7486fd
+# define AO_HAVE_test_and_set
7486fd
+
7486fd
+AO_INLINE AO_TS_VAL_t
7486fd
+AO_test_and_set_acquire(volatile AO_TS_t *addr)
7486fd
+{
7486fd
+    return (AO_TS_VAL_t)__atomic_test_and_set(addr, __ATOMIC_ACQUIRE);
7486fd
+}
7486fd
+# define AO_HAVE_test_and_set_acquire
7486fd
+
7486fd
+AO_INLINE AO_TS_VAL_t
7486fd
+AO_test_and_set_release(volatile AO_TS_t *addr)
7486fd
+{
7486fd
+    return (AO_TS_VAL_t)__atomic_test_and_set(addr, __ATOMIC_RELEASE);
7486fd
+}
7486fd
+# define AO_HAVE_test_and_set_release
7486fd
+
7486fd
+AO_INLINE AO_TS_VAL_t
7486fd
+AO_test_and_set_full(volatile AO_TS_t *addr)
7486fd
+{
7486fd
+    return (AO_TS_VAL_t)__atomic_test_and_set(addr, __ATOMIC_SEQ_CST);
7486fd
+}
7486fd
+# define AO_HAVE_test_and_set_full
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add(volatile AO_t *p, AO_t incr)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, incr, __ATOMIC_RELAXED);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add_acquire(volatile AO_t *p, AO_t incr)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, incr, __ATOMIC_ACQUIRE);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add_acquire
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add_release(volatile AO_t *p, AO_t incr)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, incr, __ATOMIC_RELEASE);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add_release
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add_full(volatile AO_t *p, AO_t incr)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, incr, __ATOMIC_SEQ_CST);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add_full
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add1(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, 1, __ATOMIC_RELAXED);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add1
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add1_acquire(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, 1, __ATOMIC_ACQUIRE);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add1_acquire
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add1_release(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, 1, __ATOMIC_RELEASE);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add1_release
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_add1_full(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_add(p, 1, __ATOMIC_SEQ_CST);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_add1_full
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_sub1(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_sub(p, 1, __ATOMIC_RELAXED);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_sub1
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_sub1_acquire(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_sub(p, 1, __ATOMIC_ACQUIRE);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_sub1_acquire
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_sub1_release(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_sub(p, 1, __ATOMIC_RELEASE);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_sub1_release
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_and_sub1_full(volatile AO_t *p)
7486fd
+{
7486fd
+  return (AO_t)__atomic_fetch_sub(p, 1, __ATOMIC_SEQ_CST);
7486fd
+}
7486fd
+#define AO_HAVE_fetch_and_sub1_full
7486fd
+
7486fd
+/* Returns nonzero if the comparison succeeded.  */
7486fd
+AO_INLINE int
7486fd
+AO_compare_and_swap(volatile AO_t *addr, AO_t old_val, AO_t new_val)
7486fd
+{
7486fd
+  return (int)__sync_bool_compare_and_swap(addr, old_val, new_val);
7486fd
+}
7486fd
+# define AO_HAVE_compare_and_swap
7486fd
+
7486fd
+AO_INLINE AO_t
7486fd
+AO_fetch_compare_and_swap(volatile AO_t *addr, AO_t old_val, AO_t new_val)
7486fd
+{
7486fd
+    return (AO_t)__sync_val_compare_and_swap(addr, old_val, new_val);
7486fd
+}
7486fd
+# define AO_HAVE_fetch_compare_and_swap
7486fd
+
7486fd
+
7486fd
+
7486fd
+#include "../../generalize.h"
7486fd
diff --git a/libatomic_ops/src/atomic_ops/sysdeps/standard_ao_double_t.h b/libatomic_ops/src/atomic_ops/sysdeps/standard_ao_double_t.h
7486fd
index 7089f05..de726fc 100644
7486fd
--- a/libatomic_ops/src/atomic_ops/sysdeps/standard_ao_double_t.h
7486fd
+++ b/libatomic_ops/src/atomic_ops/sysdeps/standard_ao_double_t.h
7486fd
@@ -11,6 +11,8 @@
7486fd
   typedef __m128 double_ptr_storage;
7486fd
 #elif defined(_WIN32) && !defined(__GNUC__)
7486fd
   typedef unsigned __int64 double_ptr_storage;
7486fd
+#elif defined(__aarch64__)
7486fd
+  typedef unsigned __int128 double_ptr_storage;
7486fd
 #else
7486fd
   typedef unsigned long long double_ptr_storage;
7486fd
 #endif