summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'dev-libs/libatomic_ops/files/libatomic_ops-7.2_alpha6-x32.patch')
-rw-r--r--dev-libs/libatomic_ops/files/libatomic_ops-7.2_alpha6-x32.patch46
1 files changed, 46 insertions, 0 deletions
diff --git a/dev-libs/libatomic_ops/files/libatomic_ops-7.2_alpha6-x32.patch b/dev-libs/libatomic_ops/files/libatomic_ops-7.2_alpha6-x32.patch
new file mode 100644
index 000000000000..838e4a8d27f1
--- /dev/null
+++ b/dev-libs/libatomic_ops/files/libatomic_ops-7.2_alpha6-x32.patch
@@ -0,0 +1,46 @@
+Adopted from
+http://git.yoctoproject.org/cgit/cgit.cgi/poky/commit/?id=e3a384578f677c05d812d99c2c92aa13670bd06a
+
+Upstream-Status: Pending
+
+Remove the `q' suffix on x86-64 atomic instructions.
+
+We don't need the `q' suffix on x86_64 atomic instructions for AO_t,
+which is defined as "unsigned long". "unsigned long" is 32bit for x32
+and 64bit for x86-64. The register operand in x86-64 atomic instructions
+is sufficient to properly determine the register size.
+
+Received this patch from H.J. Lu <hjl.tools@gmail.com>
+Signed-Off-By: Nitin A Kamble <nitin.a.kamble@intel.com> 2011/12/02
+
+diff --git a/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h b/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h
+index 0f68c1e..3bcde88 100644
+--- a/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h
++++ b/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h
+@@ -47,7 +47,7 @@ AO_fetch_and_add_full (volatile AO_t *p, AO_t incr)
+ {
+ AO_t result;
+
+- __asm__ __volatile__ ("lock; xaddq %0, %1" :
++ __asm__ __volatile__ ("lock; xadd %0, %1" :
+ "=r" (result), "=m" (*p) : "0" (incr), "m" (*p)
+ : "memory");
+ return result;
+@@ -93,7 +93,7 @@ AO_int_fetch_and_add_full (volatile unsigned int *p, unsigned int incr)
+ AO_INLINE void
+ AO_or_full (volatile AO_t *p, AO_t incr)
+ {
+- __asm__ __volatile__ ("lock; orq %1, %0" :
++ __asm__ __volatile__ ("lock; or %1, %0" :
+ "=m" (*p) : "r" (incr), "m" (*p) : "memory");
+ }
+ #define AO_HAVE_or_full
+@@ -123,7 +123,7 @@ AO_compare_and_swap_full(volatile AO_t *addr, AO_t old, AO_t new_val)
+ return (int)__sync_bool_compare_and_swap(addr, old, new_val);
+ # else
+ char result;
+- __asm__ __volatile__("lock; cmpxchgq %3, %0; setz %1"
++ __asm__ __volatile__("lock; cmpxchg %3, %0; setz %1"
+ : "=m" (*addr), "=a" (result)
+ : "m" (*addr), "r" (new_val), "a" (old) : "memory");
+ return (int) result;