1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495 |
- #ifndef _I386_PGALLOC_H
- #define _I386_PGALLOC_H
- #include <linux/threads.h>
- #include <linux/mm.h> /* for struct page */
- #include <linux/pagemap.h>
- #include <asm/tlb.h>
- #include <asm-generic/tlb.h>
- #ifdef CONFIG_PARAVIRT
- #include <asm/paravirt.h>
- #else
- #define paravirt_alloc_pt(mm, pfn) do { } while (0)
- #define paravirt_alloc_pd(mm, pfn) do { } while (0)
- #define paravirt_alloc_pd_clone(pfn, clonepfn, start, count) do { } while (0)
- #define paravirt_release_pt(pfn) do { } while (0)
- #define paravirt_release_pd(pfn) do { } while (0)
- #endif
- static inline void pmd_populate_kernel(struct mm_struct *mm,
- pmd_t *pmd, pte_t *pte)
- {
- paravirt_alloc_pt(mm, __pa(pte) >> PAGE_SHIFT);
- set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
- }
- static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, struct page *pte)
- {
- unsigned long pfn = page_to_pfn(pte);
- paravirt_alloc_pt(mm, pfn);
- set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE));
- }
- #define pmd_pgtable(pmd) pmd_page(pmd)
- /*
- * Allocate and free page tables.
- */
- extern pgd_t *pgd_alloc(struct mm_struct *);
- extern void pgd_free(struct mm_struct *mm, pgd_t *pgd);
- extern pte_t *pte_alloc_one_kernel(struct mm_struct *, unsigned long);
- extern pgtable_t pte_alloc_one(struct mm_struct *, unsigned long);
- static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
- {
- free_page((unsigned long)pte);
- }
- static inline void pte_free(struct mm_struct *mm, pgtable_t pte)
- {
- pgtable_page_dtor(pte);
- __free_page(pte);
- }
- extern void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte);
- #ifdef CONFIG_X86_PAE
- /*
- * In the PAE case we free the pmds as part of the pgd.
- */
- static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long addr)
- {
- return (pmd_t *)get_zeroed_page(GFP_KERNEL|__GFP_REPEAT);
- }
- static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd)
- {
- BUG_ON((unsigned long)pmd & (PAGE_SIZE-1));
- free_page((unsigned long)pmd);
- }
- extern void __pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd);
- static inline void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd)
- {
- paravirt_alloc_pd(mm, __pa(pmd) >> PAGE_SHIFT);
- /* Note: almost everything apart from _PAGE_PRESENT is
- reserved at the pmd (PDPT) level. */
- set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT));
- /*
- * According to Intel App note "TLBs, Paging-Structure Caches,
- * and Their Invalidation", April 2007, document 317080-001,
- * section 8.1: in PAE mode we explicitly have to flush the
- * TLB via cr3 if the top-level pgd is changed...
- */
- if (mm == current->active_mm)
- write_cr3(read_cr3());
- }
- #endif /* CONFIG_X86_PAE */
- #endif /* _I386_PGALLOC_H */
|