|
@@ -41,7 +41,7 @@ ia64_atomic_add (int i, atomic_t *v)
|
|
|
return new;
|
|
|
}
|
|
|
|
|
|
-static __inline__ int
|
|
|
+static __inline__ long
|
|
|
ia64_atomic64_add (__s64 i, atomic64_t *v)
|
|
|
{
|
|
|
__s64 old, new;
|
|
@@ -69,7 +69,7 @@ ia64_atomic_sub (int i, atomic_t *v)
|
|
|
return new;
|
|
|
}
|
|
|
|
|
|
-static __inline__ int
|
|
|
+static __inline__ long
|
|
|
ia64_atomic64_sub (__s64 i, atomic64_t *v)
|
|
|
{
|
|
|
__s64 old, new;
|
|
@@ -107,7 +107,7 @@ static __inline__ int atomic_add_unless(atomic_t *v, int a, int u)
|
|
|
|
|
|
#define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0)
|
|
|
|
|
|
-static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u)
|
|
|
+static __inline__ long atomic64_add_unless(atomic64_t *v, long a, long u)
|
|
|
{
|
|
|
long c, old;
|
|
|
c = atomic64_read(v);
|
|
@@ -158,7 +158,7 @@ atomic_add_negative (int i, atomic_t *v)
|
|
|
return atomic_add_return(i, v) < 0;
|
|
|
}
|
|
|
|
|
|
-static __inline__ int
|
|
|
+static __inline__ long
|
|
|
atomic64_add_negative (__s64 i, atomic64_t *v)
|
|
|
{
|
|
|
return atomic64_add_return(i, v) < 0;
|