123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151 |
- /* arch/sparc64/mm/tlb.c
- *
- * Copyright (C) 2004 David S. Miller <davem@redhat.com>
- */
- #include <linux/kernel.h>
- #include <linux/init.h>
- #include <linux/percpu.h>
- #include <linux/mm.h>
- #include <linux/swap.h>
- #include <asm/pgtable.h>
- #include <asm/pgalloc.h>
- #include <asm/tlbflush.h>
- #include <asm/cacheflush.h>
- #include <asm/mmu_context.h>
- #include <asm/tlb.h>
- /* Heavily inspired by the ppc64 code. */
- DEFINE_PER_CPU(struct mmu_gather, mmu_gathers) =
- { NULL, 0, 0, 0, 0, 0, { 0 }, { NULL }, };
- void flush_tlb_pending(void)
- {
- struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
- if (mp->tlb_nr) {
- if (CTX_VALID(mp->mm->context)) {
- #ifdef CONFIG_SMP
- smp_flush_tlb_pending(mp->mm, mp->tlb_nr,
- &mp->vaddrs[0]);
- #else
- __flush_tlb_pending(CTX_HWBITS(mp->mm->context),
- mp->tlb_nr, &mp->vaddrs[0]);
- #endif
- }
- mp->tlb_nr = 0;
- }
- }
- void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig)
- {
- struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
- unsigned long nr;
- vaddr &= PAGE_MASK;
- if (pte_exec(orig))
- vaddr |= 0x1UL;
- if (pte_dirty(orig)) {
- unsigned long paddr, pfn = pte_pfn(orig);
- struct address_space *mapping;
- struct page *page;
- if (!pfn_valid(pfn))
- goto no_cache_flush;
- page = pfn_to_page(pfn);
- if (PageReserved(page))
- goto no_cache_flush;
- /* A real file page? */
- mapping = page_mapping(page);
- if (!mapping)
- goto no_cache_flush;
- paddr = (unsigned long) page_address(page);
- if ((paddr ^ vaddr) & (1 << 13))
- flush_dcache_page_all(mm, page);
- }
- no_cache_flush:
- if (mp->tlb_frozen)
- return;
- nr = mp->tlb_nr;
- if (unlikely(nr != 0 && mm != mp->mm)) {
- flush_tlb_pending();
- nr = 0;
- }
- if (nr == 0)
- mp->mm = mm;
- mp->vaddrs[nr] = vaddr;
- mp->tlb_nr = ++nr;
- if (nr >= TLB_BATCH_NR)
- flush_tlb_pending();
- }
- void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end)
- {
- struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
- unsigned long nr = mp->tlb_nr;
- long s = start, e = end, vpte_base;
- if (mp->tlb_frozen)
- return;
- /* If start is greater than end, that is a real problem. */
- BUG_ON(start > end);
- /* However, straddling the VA space hole is quite normal. */
- s &= PMD_MASK;
- e = (e + PMD_SIZE - 1) & PMD_MASK;
- vpte_base = (tlb_type == spitfire ?
- VPTE_BASE_SPITFIRE :
- VPTE_BASE_CHEETAH);
- if (unlikely(nr != 0 && mm != mp->mm)) {
- flush_tlb_pending();
- nr = 0;
- }
- if (nr == 0)
- mp->mm = mm;
- start = vpte_base + (s >> (PAGE_SHIFT - 3));
- end = vpte_base + (e >> (PAGE_SHIFT - 3));
- /* If the request straddles the VA space hole, we
- * need to swap start and end. The reason this
- * occurs is that "vpte_base" is the center of
- * the linear page table mapping area. Thus,
- * high addresses with the sign bit set map to
- * addresses below vpte_base and non-sign bit
- * addresses map to addresses above vpte_base.
- */
- if (end < start) {
- unsigned long tmp = start;
- start = end;
- end = tmp;
- }
- while (start < end) {
- mp->vaddrs[nr] = start;
- mp->tlb_nr = ++nr;
- if (nr >= TLB_BATCH_NR) {
- flush_tlb_pending();
- nr = 0;
- }
- start += PAGE_SIZE;
- }
- if (nr)
- flush_tlb_pending();
- }
|