123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111 |
- #ifndef _SPARC64_TLB_H
- #define _SPARC64_TLB_H
- #include <linux/swap.h>
- #include <linux/pagemap.h>
- #include <asm/pgalloc.h>
- #include <asm/tlbflush.h>
- #include <asm/mmu_context.h>
- #define TLB_BATCH_NR 192
- /*
- * For UP we don't need to worry about TLB flush
- * and page free order so much..
- */
- #ifdef CONFIG_SMP
- #define FREE_PTE_NR 506
- #define tlb_fast_mode(bp) ((bp)->pages_nr == ~0U)
- #else
- #define FREE_PTE_NR 1
- #define tlb_fast_mode(bp) 1
- #endif
- struct mmu_gather {
- struct mm_struct *mm;
- unsigned int pages_nr;
- unsigned int need_flush;
- unsigned int fullmm;
- unsigned int tlb_nr;
- unsigned long vaddrs[TLB_BATCH_NR];
- struct page *pages[FREE_PTE_NR];
- };
- DECLARE_PER_CPU(struct mmu_gather, mmu_gathers);
- #ifdef CONFIG_SMP
- extern void smp_flush_tlb_pending(struct mm_struct *,
- unsigned long, unsigned long *);
- #endif
- extern void __flush_tlb_pending(unsigned long, unsigned long, unsigned long *);
- extern void flush_tlb_pending(void);
- static inline struct mmu_gather *tlb_gather_mmu(struct mm_struct *mm, unsigned int full_mm_flush)
- {
- struct mmu_gather *mp = &get_cpu_var(mmu_gathers);
- BUG_ON(mp->tlb_nr);
- mp->mm = mm;
- mp->pages_nr = num_online_cpus() > 1 ? 0U : ~0U;
- mp->fullmm = full_mm_flush;
- return mp;
- }
- static inline void tlb_flush_mmu(struct mmu_gather *mp)
- {
- if (!mp->fullmm)
- flush_tlb_pending();
- if (mp->need_flush) {
- free_pages_and_swap_cache(mp->pages, mp->pages_nr);
- mp->pages_nr = 0;
- mp->need_flush = 0;
- }
- }
- #ifdef CONFIG_SMP
- extern void smp_flush_tlb_mm(struct mm_struct *mm);
- #define do_flush_tlb_mm(mm) smp_flush_tlb_mm(mm)
- #else
- #define do_flush_tlb_mm(mm) __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT)
- #endif
- static inline void tlb_finish_mmu(struct mmu_gather *mp, unsigned long start, unsigned long end)
- {
- tlb_flush_mmu(mp);
- if (mp->fullmm)
- mp->fullmm = 0;
- /* keep the page table cache within bounds */
- check_pgt_cache();
- put_cpu_var(mmu_gathers);
- }
- static inline void tlb_remove_page(struct mmu_gather *mp, struct page *page)
- {
- if (tlb_fast_mode(mp)) {
- free_page_and_swap_cache(page);
- return;
- }
- mp->need_flush = 1;
- mp->pages[mp->pages_nr++] = page;
- if (mp->pages_nr >= FREE_PTE_NR)
- tlb_flush_mmu(mp);
- }
- #define tlb_remove_tlb_entry(mp,ptep,addr) do { } while (0)
- #define pte_free_tlb(mp, ptepage) pte_free((mp)->mm, ptepage)
- #define pmd_free_tlb(mp, pmdp) pmd_free((mp)->mm, pmdp)
- #define pud_free_tlb(tlb,pudp) __pud_free_tlb(tlb,pudp)
- #define tlb_migrate_finish(mm) do { } while (0)
- #define tlb_start_vma(tlb, vma) do { } while (0)
- #define tlb_end_vma(tlb, vma) do { } while (0)
- #endif /* _SPARC64_TLB_H */
|