240 lines
6.4 KiB
Diff
240 lines
6.4 KiB
Diff
From aac120d778ae5fc619b2fb8ef18ea18d3d5d20cc Mon Sep 17 00:00:00 2001
|
|
From: Yvan Roux <yvan.roux@linaro.org>
|
|
Date: Wed, 23 Jan 2013 17:14:16 +0100
|
|
Subject: [PATCH] Aarch64 basic port
|
|
|
|
Adapted-for-OpenEmbedded-by: Marcin Juszkiewicz <marcin.juszkiewicz@linaro.org>
|
|
|
|
|
|
Upstream-Status: Backport
|
|
|
|
It is original from
|
|
https://github.com/ivmai/libatomic_ops/commit/cbbf86330fcb600cfe0f895cb970d922456005d6
|
|
|
|
Signed-off-by: Kai Kang <kai.kang@windriver.com>
|
|
---
|
|
src/atomic_ops.h | 4
|
|
src/atomic_ops/sysdeps/Makefile.am | 1
|
|
src/atomic_ops/sysdeps/gcc/aarch64.h | 184 +++++++++++++++++++++++++++++++++++
|
|
3 files changed, 189 insertions(+)
|
|
create mode 100644 src/atomic_ops/sysdeps/gcc/aarch64.h
|
|
|
|
--- libatomic_ops-7.2.orig/src/atomic_ops.h
|
|
+++ libatomic_ops-7.2/src/atomic_ops.h
|
|
@@ -242,10 +242,14 @@
|
|
# endif /* __m68k__ */
|
|
# if defined(__powerpc__) || defined(__ppc__) || defined(__PPC__) \
|
|
|| defined(__powerpc64__) || defined(__ppc64__)
|
|
# include "atomic_ops/sysdeps/gcc/powerpc.h"
|
|
# endif /* __powerpc__ */
|
|
+# if defined(__aarch64__)
|
|
+# include "atomic_ops/sysdeps/gcc/aarch64.h"
|
|
+# define AO_CAN_EMUL_CAS
|
|
+# endif /* __aarch64__ */
|
|
# if defined(__arm__) && !defined(AO_USE_PTHREAD_DEFS)
|
|
# include "atomic_ops/sysdeps/gcc/arm.h"
|
|
# define AO_CAN_EMUL_CAS
|
|
# endif /* __arm__ */
|
|
# if defined(__cris__) || defined(CRIS)
|
|
--- libatomic_ops-7.2.orig/src/atomic_ops/sysdeps/Makefile.am
|
|
+++ libatomic_ops-7.2/src/atomic_ops/sysdeps/Makefile.am
|
|
@@ -24,10 +24,11 @@ nobase_sysdep_HEADERS= generic_pthread.h
|
|
standard_ao_double_t.h \
|
|
README \
|
|
\
|
|
armcc/arm_v6.h \
|
|
\
|
|
+ gcc/aarch64.h \
|
|
gcc/alpha.h gcc/arm.h gcc/avr32.h gcc/cris.h \
|
|
gcc/hexagon.h gcc/hppa.h gcc/ia64.h gcc/m68k.h \
|
|
gcc/mips.h gcc/powerpc.h gcc/s390.h \
|
|
gcc/sh.h gcc/sparc.h gcc/x86.h gcc/x86_64.h \
|
|
\
|
|
--- /dev/null
|
|
+++ libatomic_ops-7.2/src/atomic_ops/sysdeps/gcc/aarch64.h
|
|
@@ -0,0 +1,184 @@
|
|
+/*
|
|
+ * Copyright (c) 1991-1994 by Xerox Corporation. All rights reserved.
|
|
+ * Copyright (c) 1996-1999 by Silicon Graphics. All rights reserved.
|
|
+ * Copyright (c) 1999-2003 by Hewlett-Packard Company. All rights reserved.
|
|
+ *
|
|
+ *
|
|
+ * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED
|
|
+ * OR IMPLIED. ANY USE IS AT YOUR OWN RISK.
|
|
+ *
|
|
+ * Permission is hereby granted to use or copy this program
|
|
+ * for any purpose, provided the above notices are retained on all copies.
|
|
+ * Permission to modify the code and to distribute modified code is granted,
|
|
+ * provided the above notices are retained, and a notice that the code was
|
|
+ * modified is included with the above copyright notice.
|
|
+ *
|
|
+ */
|
|
+
|
|
+#include "../read_ordered.h"
|
|
+
|
|
+#include "../test_and_set_t_is_ao_t.h"
|
|
+
|
|
+AO_INLINE void
|
|
+AO_nop_full(void)
|
|
+{
|
|
+# ifndef AO_UNIPROCESSOR
|
|
+__sync_synchronize ();
|
|
+# endif
|
|
+}
|
|
+#define AO_HAVE_nop_full
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_load(const volatile AO_t *addr)
|
|
+{
|
|
+ return __atomic_load_n (addr, __ATOMIC_RELAXED);
|
|
+}
|
|
+#define AO_HAVE_load
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_load_acquire(const volatile AO_t *addr)
|
|
+{
|
|
+ return __atomic_load_n (addr, __ATOMIC_ACQUIRE);
|
|
+}
|
|
+#define AO_HAVE_load_acquire
|
|
+
|
|
+AO_INLINE void
|
|
+ AO_store(volatile AO_t *addr, AO_t value)
|
|
+{
|
|
+ __atomic_store_n(addr, value, __ATOMIC_RELAXED);
|
|
+}
|
|
+#define AO_HAVE_store
|
|
+
|
|
+AO_INLINE void
|
|
+ AO_store_release(volatile AO_t *addr, AO_t value)
|
|
+{
|
|
+ __atomic_store_n(addr, value, __ATOMIC_RELEASE);
|
|
+}
|
|
+#define AO_HAVE_store_release
|
|
+
|
|
+AO_INLINE AO_TS_VAL_t
|
|
+AO_test_and_set(volatile AO_TS_t *addr)
|
|
+{
|
|
+ return __atomic_test_and_set(addr, __ATOMIC_RELAXED);
|
|
+}
|
|
+# define AO_HAVE_test_and_set
|
|
+
|
|
+AO_INLINE AO_TS_VAL_t
|
|
+AO_test_and_set_acquire(volatile AO_TS_t *addr)
|
|
+{
|
|
+ return __atomic_test_and_set(addr, __ATOMIC_ACQUIRE);
|
|
+}
|
|
+# define AO_HAVE_test_and_set_acquire
|
|
+
|
|
+AO_INLINE AO_TS_VAL_t
|
|
+AO_test_and_set_release(volatile AO_TS_t *addr)
|
|
+{
|
|
+ return __atomic_test_and_set(addr, __ATOMIC_RELEASE);
|
|
+}
|
|
+# define AO_HAVE_test_and_set_release
|
|
+
|
|
+AO_INLINE AO_TS_VAL_t
|
|
+AO_test_and_set_full(volatile AO_TS_t *addr)
|
|
+{
|
|
+ return __atomic_test_and_set(addr, __ATOMIC_SEQ_CST);
|
|
+}
|
|
+# define AO_HAVE_test_and_set_full
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add(volatile AO_t *p, AO_t incr)
|
|
+{
|
|
+ return __atomic_fetch_add(p, incr, __ATOMIC_RELAXED);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add_acquire(volatile AO_t *p, AO_t incr)
|
|
+{
|
|
+ return __atomic_fetch_add(p, incr, __ATOMIC_ACQUIRE);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add_acquire
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add_release(volatile AO_t *p, AO_t incr)
|
|
+{
|
|
+ return __atomic_fetch_add(p, incr, __ATOMIC_RELEASE);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add_release
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add_full(volatile AO_t *p, AO_t incr)
|
|
+{
|
|
+ return __atomic_fetch_add(p, incr, __ATOMIC_SEQ_CST);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add_full
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add1(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_add(p, 1, __ATOMIC_RELAXED);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add1
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add1_acquire(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_add(p, 1, __ATOMIC_ACQUIRE);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add1_acquire
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add1_release(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_add(p, 1, __ATOMIC_RELEASE);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add1_release
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_add1_full(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_add(p, 1, __ATOMIC_SEQ_CST);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_add1_full
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_sub1(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_sub(p, 1, __ATOMIC_RELAXED);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_sub1
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_sub1_acquire(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_sub(p, 1, __ATOMIC_ACQUIRE);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_sub1_acquire
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_sub1_release(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_sub(p, 1, __ATOMIC_RELEASE);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_sub1_release
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_and_sub1_full(volatile AO_t *p)
|
|
+{
|
|
+ return __atomic_fetch_sub(p, 1, __ATOMIC_SEQ_CST);
|
|
+}
|
|
+#define AO_HAVE_fetch_and_sub1_full
|
|
+
|
|
+/* Returns nonzero if the comparison succeeded. */
|
|
+AO_INLINE int
|
|
+AO_compare_and_swap(volatile AO_t *addr, AO_t old_val, AO_t new_val)
|
|
+{
|
|
+ return __sync_bool_compare_and_swap(addr, old_val, new_val);
|
|
+}
|
|
+# define AO_HAVE_compare_and_swap
|
|
+
|
|
+AO_INLINE AO_t
|
|
+AO_fetch_compare_and_swap(volatile AO_t *addr, AO_t old_val, AO_t new_val)
|
|
+{
|
|
+ return __sync_val_compare_and_swap(addr, old_val, new_val);
|
|
+}
|
|
+# define AO_HAVE_fetch_compare_and_swap
|