tlb.c 3.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188
  1. /* arch/sparc64/mm/tlb.c
  2. *
  3. * Copyright (C) 2004 David S. Miller <davem@redhat.com>
  4. */
  5. #include <linux/kernel.h>
  6. #include <linux/init.h>
  7. #include <linux/percpu.h>
  8. #include <linux/mm.h>
  9. #include <linux/swap.h>
  10. #include <linux/preempt.h>
  11. #include <asm/pgtable.h>
  12. #include <asm/pgalloc.h>
  13. #include <asm/tlbflush.h>
  14. #include <asm/cacheflush.h>
  15. #include <asm/mmu_context.h>
  16. #include <asm/tlb.h>
  17. /* Heavily inspired by the ppc64 code. */
  18. static DEFINE_PER_CPU(struct tlb_batch, tlb_batch);
  19. void flush_tlb_pending(void)
  20. {
  21. struct tlb_batch *tb = &get_cpu_var(tlb_batch);
  22. if (tb->tlb_nr) {
  23. flush_tsb_user(tb);
  24. if (CTX_VALID(tb->mm->context)) {
  25. #ifdef CONFIG_SMP
  26. smp_flush_tlb_pending(tb->mm, tb->tlb_nr,
  27. &tb->vaddrs[0]);
  28. #else
  29. __flush_tlb_pending(CTX_HWBITS(tb->mm->context),
  30. tb->tlb_nr, &tb->vaddrs[0]);
  31. #endif
  32. }
  33. tb->tlb_nr = 0;
  34. }
  35. put_cpu_var(tlb_batch);
  36. }
  37. static void tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr,
  38. bool exec)
  39. {
  40. struct tlb_batch *tb = &get_cpu_var(tlb_batch);
  41. unsigned long nr;
  42. vaddr &= PAGE_MASK;
  43. if (exec)
  44. vaddr |= 0x1UL;
  45. nr = tb->tlb_nr;
  46. if (unlikely(nr != 0 && mm != tb->mm)) {
  47. flush_tlb_pending();
  48. nr = 0;
  49. }
  50. if (nr == 0)
  51. tb->mm = mm;
  52. tb->vaddrs[nr] = vaddr;
  53. tb->tlb_nr = ++nr;
  54. if (nr >= TLB_BATCH_NR)
  55. flush_tlb_pending();
  56. put_cpu_var(tlb_batch);
  57. }
  58. void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr,
  59. pte_t *ptep, pte_t orig, int fullmm)
  60. {
  61. if (tlb_type != hypervisor &&
  62. pte_dirty(orig)) {
  63. unsigned long paddr, pfn = pte_pfn(orig);
  64. struct address_space *mapping;
  65. struct page *page;
  66. if (!pfn_valid(pfn))
  67. goto no_cache_flush;
  68. page = pfn_to_page(pfn);
  69. if (PageReserved(page))
  70. goto no_cache_flush;
  71. /* A real file page? */
  72. mapping = page_mapping(page);
  73. if (!mapping)
  74. goto no_cache_flush;
  75. paddr = (unsigned long) page_address(page);
  76. if ((paddr ^ vaddr) & (1 << 13))
  77. flush_dcache_page_all(mm, page);
  78. }
  79. no_cache_flush:
  80. if (!fullmm)
  81. tlb_batch_add_one(mm, vaddr, pte_exec(orig));
  82. }
  83. #ifdef CONFIG_TRANSPARENT_HUGEPAGE
  84. static void tlb_batch_pmd_scan(struct mm_struct *mm, unsigned long vaddr,
  85. pmd_t pmd, bool exec)
  86. {
  87. unsigned long end;
  88. pte_t *pte;
  89. pte = pte_offset_map(&pmd, vaddr);
  90. end = vaddr + HPAGE_SIZE;
  91. while (vaddr < end) {
  92. if (pte_val(*pte) & _PAGE_VALID)
  93. tlb_batch_add_one(mm, vaddr, exec);
  94. pte++;
  95. vaddr += PAGE_SIZE;
  96. }
  97. pte_unmap(pte);
  98. }
  99. void set_pmd_at(struct mm_struct *mm, unsigned long addr,
  100. pmd_t *pmdp, pmd_t pmd)
  101. {
  102. pmd_t orig = *pmdp;
  103. *pmdp = pmd;
  104. if (mm == &init_mm)
  105. return;
  106. if ((pmd_val(pmd) ^ pmd_val(orig)) & PMD_ISHUGE) {
  107. if (pmd_val(pmd) & PMD_ISHUGE)
  108. mm->context.huge_pte_count++;
  109. else
  110. mm->context.huge_pte_count--;
  111. if (mm->context.huge_pte_count == 1)
  112. hugetlb_setup(mm);
  113. }
  114. if (!pmd_none(orig)) {
  115. bool exec = ((pmd_val(orig) & PMD_HUGE_EXEC) != 0);
  116. addr &= HPAGE_MASK;
  117. if (pmd_val(orig) & PMD_ISHUGE)
  118. tlb_batch_add_one(mm, addr, exec);
  119. else
  120. tlb_batch_pmd_scan(mm, addr, orig, exec);
  121. }
  122. }
  123. void pgtable_trans_huge_deposit(struct mm_struct *mm, pgtable_t pgtable)
  124. {
  125. struct list_head *lh = (struct list_head *) pgtable;
  126. assert_spin_locked(&mm->page_table_lock);
  127. /* FIFO */
  128. if (!mm->pmd_huge_pte)
  129. INIT_LIST_HEAD(lh);
  130. else
  131. list_add(lh, (struct list_head *) mm->pmd_huge_pte);
  132. mm->pmd_huge_pte = pgtable;
  133. }
  134. pgtable_t pgtable_trans_huge_withdraw(struct mm_struct *mm)
  135. {
  136. struct list_head *lh;
  137. pgtable_t pgtable;
  138. assert_spin_locked(&mm->page_table_lock);
  139. /* FIFO */
  140. pgtable = mm->pmd_huge_pte;
  141. lh = (struct list_head *) pgtable;
  142. if (list_empty(lh))
  143. mm->pmd_huge_pte = NULL;
  144. else {
  145. mm->pmd_huge_pte = (pgtable_t) lh->next;
  146. list_del(lh);
  147. }
  148. pte_val(pgtable[0]) = 0;
  149. pte_val(pgtable[1]) = 0;
  150. return pgtable;
  151. }
  152. #endif /* CONFIG_TRANSPARENT_HUGEPAGE */