summaryrefslogtreecommitdiff
path: root/src/atomic_ops/sysdeps/gcc/hexagon.h
diff options
context:
space:
mode:
Diffstat (limited to 'src/atomic_ops/sysdeps/gcc/hexagon.h')
-rw-r--r--src/atomic_ops/sysdeps/gcc/hexagon.h32
1 files changed, 16 insertions, 16 deletions
diff --git a/src/atomic_ops/sysdeps/gcc/hexagon.h b/src/atomic_ops/sysdeps/gcc/hexagon.h
index da7eb4e..93a236b 100644
--- a/src/atomic_ops/sysdeps/gcc/hexagon.h
+++ b/src/atomic_ops/sysdeps/gcc/hexagon.h
@@ -35,10 +35,10 @@ AO_fetch_and_add(volatile AO_t *addr, AO_t incr)
AO_t newval;
__asm__ __volatile__(
"1:\n"
- " %0 = memw_locked(%3);\n" /* load and reserve */
- " %1 = add (%0,%4);\n" /* increment */
- " memw_locked(%3,p1) = %1;\n" /* store conditional */
- " if (!p1) jump 1b;\n" /* retry if lost reservation */
+ " %0 = memw_locked(%3);\n" /* load and reserve */
+ " %1 = add (%0,%4);\n" /* increment */
+ " memw_locked(%3,p1) = %1;\n" /* store conditional */
+ " if (!p1) jump 1b;\n" /* retry if lost reservation */
: "=&r"(oldval), "=&r"(newval), "+m"(*addr)
: "r"(addr), "r"(incr)
: "memory", "p1");
@@ -54,14 +54,14 @@ AO_test_and_set(volatile AO_TS_t *addr)
__asm__ __volatile__(
"1:\n"
- " %0 = memw_locked(%2);\n" /* load and reserve */
+ " %0 = memw_locked(%2);\n" /* load and reserve */
" {\n"
- " p2 = cmp.eq(%0,#0);\n" /* if load is not zero, */
- " if (!p2.new) jump:nt 2f; \n" /* we are done */
+ " p2 = cmp.eq(%0,#0);\n" /* if load is not zero, */
+ " if (!p2.new) jump:nt 2f;\n" /* we are done */
" }\n"
- " memw_locked(%2,p1) = %3;\n" /* else store conditional */
- " if (!p1) jump 1b;\n" /* retry if lost reservation */
- "2:\n" /* oldval is zero if we set */
+ " memw_locked(%2,p1) = %3;\n" /* else store conditional */
+ " if (!p1) jump 1b;\n" /* retry if lost reservation */
+ "2:\n" /* oldval is zero if we set */
: "=&r"(oldval), "+m"(*addr)
: "r"(addr), "r"(locked_value)
: "memory", "p1", "p2");
@@ -81,7 +81,7 @@ AO_test_and_set(volatile AO_TS_t *addr)
" %0 = memw_locked(%3);\n" /* load and reserve */
" {\n"
" p2 = cmp.eq(%0,%4);\n" /* if load is not equal to */
- " if (!p2.new) jump:nt 2f; \n" /* old, fail */
+ " if (!p2.new) jump:nt 2f;\n" /* old, fail */
" }\n"
" memw_locked(%3,p1) = %5;\n" /* else store conditional */
" if (!p1) jump 1b;\n" /* retry if lost reservation */
@@ -103,13 +103,13 @@ AO_fetch_compare_and_swap(volatile AO_t *addr, AO_t old_val, AO_t new_val)
__asm__ __volatile__(
"1:\n"
- " %0 = memw_locked(%2);\n" /* load and reserve */
+ " %0 = memw_locked(%2);\n" /* load and reserve */
" {\n"
- " p2 = cmp.eq(%0,%3);\n" /* if load is not equal to */
- " if (!p2.new) jump:nt 2f; \n" /* old_val, fail */
+ " p2 = cmp.eq(%0,%3);\n" /* if load is not equal to */
+ " if (!p2.new) jump:nt 2f;\n" /* old_val, fail */
" }\n"
- " memw_locked(%2,p1) = %4;\n" /* else store conditional */
- " if (!p1) jump 1b;\n" /* retry if lost reservation */
+ " memw_locked(%2,p1) = %4;\n" /* else store conditional */
+ " if (!p1) jump 1b;\n" /* retry if lost reservation */
"2:\n"
: "=&r" (__oldval), "+m"(*addr)
: "r" (addr), "r" (old_val), "r" (new_val)