tlb.h 2.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128
  1. #ifndef _SPARC64_TLB_H
  2. #define _SPARC64_TLB_H
  3. #include <linux/config.h>
  4. #include <linux/swap.h>
  5. #include <asm/pgalloc.h>
  6. #include <asm/tlbflush.h>
  7. #include <asm/mmu_context.h>
  8. #define TLB_BATCH_NR 192
  9. /*
  10. * For UP we don't need to worry about TLB flush
  11. * and page free order so much..
  12. */
  13. #ifdef CONFIG_SMP
  14. #define FREE_PTE_NR 506
  15. #define tlb_fast_mode(bp) ((bp)->pages_nr == ~0U)
  16. #else
  17. #define FREE_PTE_NR 1
  18. #define tlb_fast_mode(bp) 1
  19. #endif
  20. struct mmu_gather {
  21. struct mm_struct *mm;
  22. unsigned int pages_nr;
  23. unsigned int need_flush;
  24. unsigned int tlb_frozen;
  25. unsigned int tlb_nr;
  26. unsigned long freed;
  27. unsigned long vaddrs[TLB_BATCH_NR];
  28. struct page *pages[FREE_PTE_NR];
  29. };
  30. DECLARE_PER_CPU(struct mmu_gather, mmu_gathers);
  31. #ifdef CONFIG_SMP
  32. extern void smp_flush_tlb_pending(struct mm_struct *,
  33. unsigned long, unsigned long *);
  34. #endif
  35. extern void __flush_tlb_pending(unsigned long, unsigned long, unsigned long *);
  36. extern void flush_tlb_pending(void);
  37. static inline struct mmu_gather *tlb_gather_mmu(struct mm_struct *mm, unsigned int full_mm_flush)
  38. {
  39. struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
  40. BUG_ON(mp->tlb_nr);
  41. mp->mm = mm;
  42. mp->pages_nr = num_online_cpus() > 1 ? 0U : ~0U;
  43. mp->tlb_frozen = full_mm_flush;
  44. mp->freed = 0;
  45. return mp;
  46. }
  47. static inline void tlb_flush_mmu(struct mmu_gather *mp)
  48. {
  49. if (mp->need_flush) {
  50. mp->need_flush = 0;
  51. if (!tlb_fast_mode(mp)) {
  52. free_pages_and_swap_cache(mp->pages, mp->pages_nr);
  53. mp->pages_nr = 0;
  54. }
  55. }
  56. }
  57. #ifdef CONFIG_SMP
  58. extern void smp_flush_tlb_mm(struct mm_struct *mm);
  59. #define do_flush_tlb_mm(mm) smp_flush_tlb_mm(mm)
  60. #else
  61. #define do_flush_tlb_mm(mm) __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT)
  62. #endif
  63. static inline void tlb_finish_mmu(struct mmu_gather *mp, unsigned long start, unsigned long end)
  64. {
  65. unsigned long freed = mp->freed;
  66. struct mm_struct *mm = mp->mm;
  67. unsigned long rss = get_mm_counter(mm, rss);
  68. if (rss < freed)
  69. freed = rss;
  70. add_mm_counter(mm, rss, -freed);
  71. tlb_flush_mmu(mp);
  72. if (mp->tlb_frozen) {
  73. if (CTX_VALID(mm->context))
  74. do_flush_tlb_mm(mm);
  75. mp->tlb_frozen = 0;
  76. } else
  77. flush_tlb_pending();
  78. /* keep the page table cache within bounds */
  79. check_pgt_cache();
  80. }
  81. static inline unsigned int tlb_is_full_mm(struct mmu_gather *mp)
  82. {
  83. return mp->tlb_frozen;
  84. }
  85. static inline void tlb_remove_page(struct mmu_gather *mp, struct page *page)
  86. {
  87. mp->need_flush = 1;
  88. if (tlb_fast_mode(mp)) {
  89. free_page_and_swap_cache(page);
  90. return;
  91. }
  92. mp->pages[mp->pages_nr++] = page;
  93. if (mp->pages_nr >= FREE_PTE_NR)
  94. tlb_flush_mmu(mp);
  95. }
  96. #define tlb_remove_tlb_entry(mp,ptep,addr) do { } while (0)
  97. #define pte_free_tlb(mp,ptepage) pte_free(ptepage)
  98. #define pmd_free_tlb(mp,pmdp) pmd_free(pmdp)
  99. #define pud_free_tlb(tlb,pudp) __pud_free_tlb(tlb,pudp)
  100. #define tlb_migrate_finish(mm) do { } while (0)
  101. #define tlb_start_vma(tlb, vma) do { } while (0)
  102. #define tlb_end_vma(tlb, vma) do { } while (0)
  103. #endif /* _SPARC64_TLB_H */