|
@@ -314,16 +314,31 @@ static void __update_mmu_tsb_insert(struct mm_struct *mm, unsigned long tsb_inde
|
|
|
struct tsb *tsb = mm->context.tsb_block[tsb_index].tsb;
|
|
|
unsigned long tag;
|
|
|
|
|
|
+ if (unlikely(!tsb))
|
|
|
+ return;
|
|
|
+
|
|
|
tsb += ((address >> tsb_hash_shift) &
|
|
|
(mm->context.tsb_block[tsb_index].tsb_nentries - 1UL));
|
|
|
tag = (address >> 22UL);
|
|
|
tsb_insert(tsb, tag, tte);
|
|
|
}
|
|
|
|
|
|
+#if defined(CONFIG_HUGETLB_PAGE) || defined(CONFIG_TRANSPARENT_HUGEPAGE)
|
|
|
+static inline bool is_hugetlb_pte(pte_t pte)
|
|
|
+{
|
|
|
+ if ((tlb_type == hypervisor &&
|
|
|
+ (pte_val(pte) & _PAGE_SZALL_4V) == _PAGE_SZHUGE_4V) ||
|
|
|
+ (tlb_type != hypervisor &&
|
|
|
+ (pte_val(pte) & _PAGE_SZALL_4U) == _PAGE_SZHUGE_4U))
|
|
|
+ return true;
|
|
|
+ return false;
|
|
|
+}
|
|
|
+#endif
|
|
|
+
|
|
|
void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep)
|
|
|
{
|
|
|
- unsigned long tsb_index, tsb_hash_shift, flags;
|
|
|
struct mm_struct *mm;
|
|
|
+ unsigned long flags;
|
|
|
pte_t pte = *ptep;
|
|
|
|
|
|
if (tlb_type != hypervisor) {
|
|
@@ -335,25 +350,16 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *
|
|
|
|
|
|
mm = vma->vm_mm;
|
|
|
|
|
|
- tsb_index = MM_TSB_BASE;
|
|
|
- tsb_hash_shift = PAGE_SHIFT;
|
|
|
-
|
|
|
spin_lock_irqsave(&mm->context.lock, flags);
|
|
|
|
|
|
#if defined(CONFIG_HUGETLB_PAGE) || defined(CONFIG_TRANSPARENT_HUGEPAGE)
|
|
|
- if (mm->context.tsb_block[MM_TSB_HUGE].tsb != NULL) {
|
|
|
- if ((tlb_type == hypervisor &&
|
|
|
- (pte_val(pte) & _PAGE_SZALL_4V) == _PAGE_SZHUGE_4V) ||
|
|
|
- (tlb_type != hypervisor &&
|
|
|
- (pte_val(pte) & _PAGE_SZALL_4U) == _PAGE_SZHUGE_4U)) {
|
|
|
- tsb_index = MM_TSB_HUGE;
|
|
|
- tsb_hash_shift = HPAGE_SHIFT;
|
|
|
- }
|
|
|
- }
|
|
|
+ if (mm->context.huge_pte_count && is_hugetlb_pte(pte))
|
|
|
+ __update_mmu_tsb_insert(mm, MM_TSB_HUGE, HPAGE_SHIFT,
|
|
|
+ address, pte_val(pte));
|
|
|
+ else
|
|
|
#endif
|
|
|
-
|
|
|
- __update_mmu_tsb_insert(mm, tsb_index, tsb_hash_shift,
|
|
|
- address, pte_val(pte));
|
|
|
+ __update_mmu_tsb_insert(mm, MM_TSB_BASE, PAGE_SHIFT,
|
|
|
+ address, pte_val(pte));
|
|
|
|
|
|
spin_unlock_irqrestore(&mm->context.lock, flags);
|
|
|
}
|
|
@@ -2712,14 +2718,28 @@ static void context_reload(void *__data)
|
|
|
load_secondary_context(mm);
|
|
|
}
|
|
|
|
|
|
-void hugetlb_setup(struct mm_struct *mm)
|
|
|
+void hugetlb_setup(struct pt_regs *regs)
|
|
|
{
|
|
|
- struct tsb_config *tp = &mm->context.tsb_block[MM_TSB_HUGE];
|
|
|
+ struct mm_struct *mm = current->mm;
|
|
|
+ struct tsb_config *tp;
|
|
|
|
|
|
- if (likely(tp->tsb != NULL))
|
|
|
- return;
|
|
|
+ if (in_atomic() || !mm) {
|
|
|
+ const struct exception_table_entry *entry;
|
|
|
+
|
|
|
+ entry = search_exception_tables(regs->tpc);
|
|
|
+ if (entry) {
|
|
|
+ regs->tpc = entry->fixup;
|
|
|
+ regs->tnpc = regs->tpc + 4;
|
|
|
+ return;
|
|
|
+ }
|
|
|
+ pr_alert("Unexpected HugeTLB setup in atomic context.\n");
|
|
|
+ die_if_kernel("HugeTSB in atomic", regs);
|
|
|
+ }
|
|
|
+
|
|
|
+ tp = &mm->context.tsb_block[MM_TSB_HUGE];
|
|
|
+ if (likely(tp->tsb == NULL))
|
|
|
+ tsb_grow(mm, MM_TSB_HUGE, 0);
|
|
|
|
|
|
- tsb_grow(mm, MM_TSB_HUGE, 0);
|
|
|
tsb_context_switch(mm);
|
|
|
smp_tsb_sync(mm);
|
|
|
|