Browse Source

um: merge arch/um/sys-{i386,x86_64}/shared

Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
Signed-off-by: Richard Weinberger <richard@nod.at>
Al Viro 14 years ago
parent
commit
858ba94499
39 changed files with 61 additions and 134 deletions
  1. 1 2
      arch/um/Makefile
  2. 5 0
      arch/um/sys-x86/shared/sysdep/archsetjmp.h
  3. 0 0
      arch/um/sys-x86/shared/sysdep/archsetjmp_32.h
  4. 0 0
      arch/um/sys-x86/shared/sysdep/archsetjmp_64.h
  5. 5 0
      arch/um/sys-x86/shared/sysdep/barrier.h
  6. 0 0
      arch/um/sys-x86/shared/sysdep/barrier_32.h
  7. 0 0
      arch/um/sys-x86/shared/sysdep/barrier_64.h
  8. 5 0
      arch/um/sys-x86/shared/sysdep/faultinfo.h
  9. 0 0
      arch/um/sys-x86/shared/sysdep/faultinfo_32.h
  10. 0 0
      arch/um/sys-x86/shared/sysdep/faultinfo_64.h
  11. 5 0
      arch/um/sys-x86/shared/sysdep/host_ldt.h
  12. 0 0
      arch/um/sys-x86/shared/sysdep/host_ldt_32.h
  13. 0 0
      arch/um/sys-x86/shared/sysdep/host_ldt_64.h
  14. 5 0
      arch/um/sys-x86/shared/sysdep/kernel-offsets.h
  15. 0 0
      arch/um/sys-x86/shared/sysdep/kernel-offsets_32.h
  16. 0 0
      arch/um/sys-x86/shared/sysdep/kernel-offsets_64.h
  17. 5 0
      arch/um/sys-x86/shared/sysdep/ptrace.h
  18. 0 0
      arch/um/sys-x86/shared/sysdep/ptrace_32.h
  19. 0 0
      arch/um/sys-x86/shared/sysdep/ptrace_64.h
  20. 5 0
      arch/um/sys-x86/shared/sysdep/ptrace_user.h
  21. 0 0
      arch/um/sys-x86/shared/sysdep/ptrace_user_32.h
  22. 0 0
      arch/um/sys-x86/shared/sysdep/ptrace_user_64.h
  23. 5 0
      arch/um/sys-x86/shared/sysdep/sigcontext.h
  24. 0 0
      arch/um/sys-x86/shared/sysdep/sigcontext_32.h
  25. 0 0
      arch/um/sys-x86/shared/sysdep/sigcontext_64.h
  26. 5 0
      arch/um/sys-x86/shared/sysdep/skas_ptrace.h
  27. 0 0
      arch/um/sys-x86/shared/sysdep/skas_ptrace_32.h
  28. 0 0
      arch/um/sys-x86/shared/sysdep/skas_ptrace_64.h
  29. 5 0
      arch/um/sys-x86/shared/sysdep/stub.h
  30. 0 0
      arch/um/sys-x86/shared/sysdep/stub_32.h
  31. 0 0
      arch/um/sys-x86/shared/sysdep/stub_64.h
  32. 5 0
      arch/um/sys-x86/shared/sysdep/syscalls.h
  33. 0 0
      arch/um/sys-x86/shared/sysdep/syscalls_32.h
  34. 0 0
      arch/um/sys-x86/shared/sysdep/syscalls_64.h
  35. 0 0
      arch/um/sys-x86/shared/sysdep/system.h
  36. 5 0
      arch/um/sys-x86/shared/sysdep/tls.h
  37. 0 0
      arch/um/sys-x86/shared/sysdep/tls_32.h
  38. 0 0
      arch/um/sys-x86/shared/sysdep/tls_64.h
  39. 0 132
      arch/um/sys-x86_64/shared/sysdep/system.h

+ 1 - 2
arch/um/Makefile

@@ -25,9 +25,8 @@ include $(srctree)/$(ARCH_DIR)/Makefile-$(SUBARCH)
 
 SHARED_HEADERS	:= $(ARCH_DIR)/include/shared
 ARCH_INCLUDE	:= -I$(srctree)/$(SHARED_HEADERS)
-ARCH_INCLUDE	+= -I$(srctree)/$(ARCH_DIR)/sys-$(SUBARCH)/shared
+ARCH_INCLUDE	+= -I$(srctree)/$(ARCH_DIR)/sys-$(HEADER_ARCH)/shared
 KBUILD_CPPFLAGS += -I$(srctree)/$(ARCH_DIR)/sys-$(HEADER_ARCH)
-KBUILD_CPPFLAGS += -I$(srctree)/$(ARCH_DIR)/sys-$(SUBARCH)
 
 # -Dvmap=kernel_vmap prevents anything from referencing the libpcap.o symbol so
 # named - it's a common symbol in libpcap, so we get a binary which crashes.

+ 5 - 0
arch/um/sys-x86/shared/sysdep/archsetjmp.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "archsetjmp_32.h"
+#else
+#include "archsetjmp_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/archsetjmp.h → arch/um/sys-x86/shared/sysdep/archsetjmp_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/archsetjmp.h → arch/um/sys-x86/shared/sysdep/archsetjmp_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/barrier.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "barrier_32.h"
+#else
+#include "barrier_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/barrier.h → arch/um/sys-x86/shared/sysdep/barrier_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/barrier.h → arch/um/sys-x86/shared/sysdep/barrier_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/faultinfo.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "faultinfo_32.h"
+#else
+#include "faultinfo_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/faultinfo.h → arch/um/sys-x86/shared/sysdep/faultinfo_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/faultinfo.h → arch/um/sys-x86/shared/sysdep/faultinfo_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/host_ldt.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "host_ldt_32.h"
+#else
+#include "host_ldt_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/host_ldt.h → arch/um/sys-x86/shared/sysdep/host_ldt_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/host_ldt.h → arch/um/sys-x86/shared/sysdep/host_ldt_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/kernel-offsets.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "kernel-offsets_32.h"
+#else
+#include "kernel-offsets_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/kernel-offsets.h → arch/um/sys-x86/shared/sysdep/kernel-offsets_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/kernel-offsets.h → arch/um/sys-x86/shared/sysdep/kernel-offsets_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/ptrace.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "ptrace_32.h"
+#else
+#include "ptrace_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/ptrace.h → arch/um/sys-x86/shared/sysdep/ptrace_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/ptrace.h → arch/um/sys-x86/shared/sysdep/ptrace_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/ptrace_user.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "ptrace_user_32.h"
+#else
+#include "ptrace_user_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/ptrace_user.h → arch/um/sys-x86/shared/sysdep/ptrace_user_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/ptrace_user.h → arch/um/sys-x86/shared/sysdep/ptrace_user_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/sigcontext.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "sigcontext_32.h"
+#else
+#include "sigcontext_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/sigcontext.h → arch/um/sys-x86/shared/sysdep/sigcontext_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/sigcontext.h → arch/um/sys-x86/shared/sysdep/sigcontext_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/skas_ptrace.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "skas_ptrace_32.h"
+#else
+#include "skas_ptrace_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/skas_ptrace.h → arch/um/sys-x86/shared/sysdep/skas_ptrace_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/skas_ptrace.h → arch/um/sys-x86/shared/sysdep/skas_ptrace_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/stub.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "stub_32.h"
+#else
+#include "stub_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/stub.h → arch/um/sys-x86/shared/sysdep/stub_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/stub.h → arch/um/sys-x86/shared/sysdep/stub_64.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/syscalls.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "syscalls_32.h"
+#else
+#include "syscalls_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/syscalls.h → arch/um/sys-x86/shared/sysdep/syscalls_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/syscalls.h → arch/um/sys-x86/shared/sysdep/syscalls_64.h


+ 0 - 0
arch/um/sys-i386/shared/sysdep/system.h → arch/um/sys-x86/shared/sysdep/system.h


+ 5 - 0
arch/um/sys-x86/shared/sysdep/tls.h

@@ -0,0 +1,5 @@
+#ifdef __i386__
+#include "tls_32.h"
+#else
+#include "tls_64.h"
+#endif

+ 0 - 0
arch/um/sys-i386/shared/sysdep/tls.h → arch/um/sys-x86/shared/sysdep/tls_32.h


+ 0 - 0
arch/um/sys-x86_64/shared/sysdep/tls.h → arch/um/sys-x86/shared/sysdep/tls_64.h


+ 0 - 132
arch/um/sys-x86_64/shared/sysdep/system.h

@@ -1,132 +0,0 @@
-#ifndef _ASM_X86_SYSTEM_H_
-#define _ASM_X86_SYSTEM_H_
-
-#include <asm/asm.h>
-#include <asm/segment.h>
-#include <asm/cpufeature.h>
-#include <asm/cmpxchg.h>
-#include <asm/nops.h>
-
-#include <linux/kernel.h>
-#include <linux/irqflags.h>
-
-/* entries in ARCH_DLINFO: */
-#ifdef CONFIG_IA32_EMULATION
-# define AT_VECTOR_SIZE_ARCH 2
-#else
-# define AT_VECTOR_SIZE_ARCH 1
-#endif
-
-extern unsigned long arch_align_stack(unsigned long sp);
-
-void default_idle(void);
-
-/*
- * Force strict CPU ordering.
- * And yes, this is required on UP too when we're talking
- * to devices.
- */
-#ifdef CONFIG_X86_32
-/*
- * Some non-Intel clones support out of order store. wmb() ceases to be a
- * nop for these.
- */
-#define mb() alternative("lock; addl $0,0(%%esp)", "mfence", X86_FEATURE_XMM2)
-#define rmb() alternative("lock; addl $0,0(%%esp)", "lfence", X86_FEATURE_XMM2)
-#define wmb() alternative("lock; addl $0,0(%%esp)", "sfence", X86_FEATURE_XMM)
-#else
-#define mb() 	asm volatile("mfence":::"memory")
-#define rmb()	asm volatile("lfence":::"memory")
-#define wmb()	asm volatile("sfence" ::: "memory")
-#endif
-
-/**
- * read_barrier_depends - Flush all pending reads that subsequents reads
- * depend on.
- *
- * No data-dependent reads from memory-like regions are ever reordered
- * over this barrier.  All reads preceding this primitive are guaranteed
- * to access memory (but not necessarily other CPUs' caches) before any
- * reads following this primitive that depend on the data return by
- * any of the preceding reads.  This primitive is much lighter weight than
- * rmb() on most CPUs, and is never heavier weight than is
- * rmb().
- *
- * These ordering constraints are respected by both the local CPU
- * and the compiler.
- *
- * Ordering is not guaranteed by anything other than these primitives,
- * not even by data dependencies.  See the documentation for
- * memory_barrier() for examples and URLs to more information.
- *
- * For example, the following code would force ordering (the initial
- * value of "a" is zero, "b" is one, and "p" is "&a"):
- *
- * <programlisting>
- *	CPU 0				CPU 1
- *
- *	b = 2;
- *	memory_barrier();
- *	p = &b;				q = p;
- *					read_barrier_depends();
- *					d = *q;
- * </programlisting>
- *
- * because the read of "*q" depends on the read of "p" and these
- * two reads are separated by a read_barrier_depends().  However,
- * the following code, with the same initial values for "a" and "b":
- *
- * <programlisting>
- *	CPU 0				CPU 1
- *
- *	a = 2;
- *	memory_barrier();
- *	b = 3;				y = b;
- *					read_barrier_depends();
- *					x = a;
- * </programlisting>
- *
- * does not enforce ordering, since there is no data dependency between
- * the read of "a" and the read of "b".  Therefore, on some CPUs, such
- * as Alpha, "y" could be set to 3 and "x" to 0.  Use rmb()
- * in cases like this where there are no data dependencies.
- **/
-
-#define read_barrier_depends()	do { } while (0)
-
-#ifdef CONFIG_SMP
-#define smp_mb()	mb()
-#ifdef CONFIG_X86_PPRO_FENCE
-# define smp_rmb()	rmb()
-#else
-# define smp_rmb()	barrier()
-#endif
-#ifdef CONFIG_X86_OOSTORE
-# define smp_wmb() 	wmb()
-#else
-# define smp_wmb()	barrier()
-#endif
-#define smp_read_barrier_depends()	read_barrier_depends()
-#define set_mb(var, value) do { (void)xchg(&var, value); } while (0)
-#else
-#define smp_mb()	barrier()
-#define smp_rmb()	barrier()
-#define smp_wmb()	barrier()
-#define smp_read_barrier_depends()	do { } while (0)
-#define set_mb(var, value) do { var = value; barrier(); } while (0)
-#endif
-
-/*
- * Stop RDTSC speculation. This is needed when you need to use RDTSC
- * (or get_cycles or vread that possibly accesses the TSC) in a defined
- * code region.
- *
- * (Could use an alternative three way for this if there was one.)
- */
-static inline void rdtsc_barrier(void)
-{
-	alternative(ASM_NOP3, "mfence", X86_FEATURE_MFENCE_RDTSC);
-	alternative(ASM_NOP3, "lfence", X86_FEATURE_LFENCE_RDTSC);
-}
-
-#endif