|
@@ -216,6 +216,7 @@ static inline unsigned long get_limit(unsigned long segment)
|
|
|
|
|
|
#define mb() alternative("lock; addl $0,0(%%esp)", "mfence", X86_FEATURE_XMM2)
|
|
#define mb() alternative("lock; addl $0,0(%%esp)", "mfence", X86_FEATURE_XMM2)
|
|
#define rmb() alternative("lock; addl $0,0(%%esp)", "lfence", X86_FEATURE_XMM2)
|
|
#define rmb() alternative("lock; addl $0,0(%%esp)", "lfence", X86_FEATURE_XMM2)
|
|
|
|
+#define wmb() alternative("lock; addl $0,0(%%esp)", "sfence", X86_FEATURE_XMM)
|
|
|
|
|
|
/**
|
|
/**
|
|
* read_barrier_depends - Flush all pending reads that subsequents reads
|
|
* read_barrier_depends - Flush all pending reads that subsequents reads
|
|
@@ -271,18 +272,14 @@ static inline unsigned long get_limit(unsigned long segment)
|
|
|
|
|
|
#define read_barrier_depends() do { } while(0)
|
|
#define read_barrier_depends() do { } while(0)
|
|
|
|
|
|
-#ifdef CONFIG_X86_OOSTORE
|
|
|
|
-/* Actually there are no OOO store capable CPUs for now that do SSE,
|
|
|
|
- but make it already an possibility. */
|
|
|
|
-#define wmb() alternative("lock; addl $0,0(%%esp)", "sfence", X86_FEATURE_XMM)
|
|
|
|
-#else
|
|
|
|
-#define wmb() __asm__ __volatile__ ("": : :"memory")
|
|
|
|
-#endif
|
|
|
|
-
|
|
|
|
#ifdef CONFIG_SMP
|
|
#ifdef CONFIG_SMP
|
|
#define smp_mb() mb()
|
|
#define smp_mb() mb()
|
|
#define smp_rmb() rmb()
|
|
#define smp_rmb() rmb()
|
|
-#define smp_wmb() wmb()
|
|
|
|
|
|
+#ifdef CONFIG_X86_OOSTORE
|
|
|
|
+# define smp_wmb() wmb()
|
|
|
|
+#else
|
|
|
|
+# define smp_wmb() barrier()
|
|
|
|
+#endif
|
|
#define smp_read_barrier_depends() read_barrier_depends()
|
|
#define smp_read_barrier_depends() read_barrier_depends()
|
|
#define set_mb(var, value) do { (void) xchg(&var, value); } while (0)
|
|
#define set_mb(var, value) do { (void) xchg(&var, value); } while (0)
|
|
#else
|
|
#else
|