|
@@ -21,17 +21,6 @@
|
|
|
#define PGTABLE_RANGE (ASM_CONST(1) << PGTABLE_EADDR_SIZE)
|
|
|
|
|
|
|
|
|
-/* Some sanity checking */
|
|
|
-#if TASK_SIZE_USER64 > PGTABLE_RANGE
|
|
|
-#error TASK_SIZE_USER64 exceeds pagetable range
|
|
|
-#endif
|
|
|
-
|
|
|
-#ifdef CONFIG_PPC_STD_MMU_64
|
|
|
-#if TASK_SIZE_USER64 > (1UL << (USER_ESID_BITS + SID_SHIFT))
|
|
|
-#error TASK_SIZE_USER64 exceeds user VSID range
|
|
|
-#endif
|
|
|
-#endif
|
|
|
-
|
|
|
/*
|
|
|
* Define the address range of the kernel non-linear virtual area
|
|
|
*/
|
|
@@ -117,9 +106,6 @@
|
|
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
|
|
-#include <linux/stddef.h>
|
|
|
-#include <asm/tlbflush.h>
|
|
|
-
|
|
|
/*
|
|
|
* This is the default implementation of various PTE accessors, it's
|
|
|
* used in all cases except Book3S with 64K pages where we have a
|
|
@@ -198,7 +184,8 @@
|
|
|
/* to find an entry in a kernel page-table-directory */
|
|
|
/* This now only contains the vmalloc pages */
|
|
|
#define pgd_offset_k(address) pgd_offset(&init_mm, address)
|
|
|
-
|
|
|
+extern void hpte_need_flush(struct mm_struct *mm, unsigned long addr,
|
|
|
+ pte_t *ptep, unsigned long pte, int huge);
|
|
|
|
|
|
/* Atomic PTE updates */
|
|
|
static inline unsigned long pte_update(struct mm_struct *mm,
|