|
@@ -59,8 +59,8 @@ static __inline__ void atomic_add(int i, atomic_t * v)
|
|
|
" sc %0, %1 \n"
|
|
|
" beqzl %0, 1b \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else if (kernel_uses_llsc) {
|
|
|
int temp;
|
|
|
|
|
@@ -71,8 +71,8 @@ static __inline__ void atomic_add(int i, atomic_t * v)
|
|
|
" addu %0, %2 \n"
|
|
|
" sc %0, %1 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} while (unlikely(!temp));
|
|
|
} else {
|
|
|
unsigned long flags;
|
|
@@ -102,8 +102,8 @@ static __inline__ void atomic_sub(int i, atomic_t * v)
|
|
|
" sc %0, %1 \n"
|
|
|
" beqzl %0, 1b \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else if (kernel_uses_llsc) {
|
|
|
int temp;
|
|
|
|
|
@@ -114,8 +114,8 @@ static __inline__ void atomic_sub(int i, atomic_t * v)
|
|
|
" subu %0, %2 \n"
|
|
|
" sc %0, %1 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} while (unlikely(!temp));
|
|
|
} else {
|
|
|
unsigned long flags;
|
|
@@ -146,9 +146,8 @@ static __inline__ int atomic_add_return(int i, atomic_t * v)
|
|
|
" beqzl %0, 1b \n"
|
|
|
" addu %0, %1, %3 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter)
|
|
|
- : "memory");
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else if (kernel_uses_llsc) {
|
|
|
int temp;
|
|
|
|
|
@@ -159,9 +158,8 @@ static __inline__ int atomic_add_return(int i, atomic_t * v)
|
|
|
" addu %0, %1, %3 \n"
|
|
|
" sc %0, %2 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter)
|
|
|
- : "memory");
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} while (unlikely(!result));
|
|
|
|
|
|
result = temp + i;
|
|
@@ -212,9 +210,8 @@ static __inline__ int atomic_sub_return(int i, atomic_t * v)
|
|
|
" subu %0, %1, %3 \n"
|
|
|
" sc %0, %2 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter)
|
|
|
- : "memory");
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} while (unlikely(!result));
|
|
|
|
|
|
result = temp - i;
|
|
@@ -262,7 +259,7 @@ static __inline__ int atomic_sub_if_positive(int i, atomic_t * v)
|
|
|
" .set reorder \n"
|
|
|
"1: \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
: "Ir" (i), "m" (v->counter)
|
|
|
: "memory");
|
|
|
} else if (kernel_uses_llsc) {
|
|
@@ -280,9 +277,8 @@ static __inline__ int atomic_sub_if_positive(int i, atomic_t * v)
|
|
|
" .set reorder \n"
|
|
|
"1: \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter)
|
|
|
- : "memory");
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else {
|
|
|
unsigned long flags;
|
|
|
|
|
@@ -430,8 +426,8 @@ static __inline__ void atomic64_add(long i, atomic64_t * v)
|
|
|
" scd %0, %1 \n"
|
|
|
" beqzl %0, 1b \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else if (kernel_uses_llsc) {
|
|
|
long temp;
|
|
|
|
|
@@ -442,8 +438,8 @@ static __inline__ void atomic64_add(long i, atomic64_t * v)
|
|
|
" daddu %0, %2 \n"
|
|
|
" scd %0, %1 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} while (unlikely(!temp));
|
|
|
} else {
|
|
|
unsigned long flags;
|
|
@@ -473,8 +469,8 @@ static __inline__ void atomic64_sub(long i, atomic64_t * v)
|
|
|
" scd %0, %1 \n"
|
|
|
" beqzl %0, 1b \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else if (kernel_uses_llsc) {
|
|
|
long temp;
|
|
|
|
|
@@ -485,8 +481,8 @@ static __inline__ void atomic64_sub(long i, atomic64_t * v)
|
|
|
" dsubu %0, %2 \n"
|
|
|
" scd %0, %1 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter));
|
|
|
+ : "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} while (unlikely(!temp));
|
|
|
} else {
|
|
|
unsigned long flags;
|
|
@@ -517,9 +513,8 @@ static __inline__ long atomic64_add_return(long i, atomic64_t * v)
|
|
|
" beqzl %0, 1b \n"
|
|
|
" daddu %0, %1, %3 \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter)
|
|
|
- : "memory");
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else if (kernel_uses_llsc) {
|
|
|
long temp;
|
|
|
|
|
@@ -649,9 +644,8 @@ static __inline__ long atomic64_sub_if_positive(long i, atomic64_t * v)
|
|
|
" .set reorder \n"
|
|
|
"1: \n"
|
|
|
" .set mips0 \n"
|
|
|
- : "=&r" (result), "=&r" (temp), "=m" (v->counter)
|
|
|
- : "Ir" (i), "m" (v->counter)
|
|
|
- : "memory");
|
|
|
+ : "=&r" (result), "=&r" (temp), "+m" (v->counter)
|
|
|
+ : "Ir" (i));
|
|
|
} else {
|
|
|
unsigned long flags;
|
|
|
|