hugetlb.h 4.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184
  1. /*
  2. * IBM System z Huge TLB Page Support for Kernel.
  3. *
  4. * Copyright IBM Corp. 2008
  5. * Author(s): Gerald Schaefer <gerald.schaefer@de.ibm.com>
  6. */
  7. #ifndef _ASM_S390_HUGETLB_H
  8. #define _ASM_S390_HUGETLB_H
  9. #include <asm/page.h>
  10. #include <asm/pgtable.h>
  11. #define is_hugepage_only_range(mm, addr, len) 0
  12. #define hugetlb_free_pgd_range free_pgd_range
  13. void set_huge_pte_at(struct mm_struct *mm, unsigned long addr,
  14. pte_t *ptep, pte_t pte);
  15. /*
  16. * If the arch doesn't supply something else, assume that hugepage
  17. * size aligned regions are ok without further preparation.
  18. */
  19. static inline int prepare_hugepage_range(struct file *file,
  20. unsigned long addr, unsigned long len)
  21. {
  22. if (len & ~HPAGE_MASK)
  23. return -EINVAL;
  24. if (addr & ~HPAGE_MASK)
  25. return -EINVAL;
  26. return 0;
  27. }
  28. #define hugetlb_prefault_arch_hook(mm) do { } while (0)
  29. int arch_prepare_hugepage(struct page *page);
  30. void arch_release_hugepage(struct page *page);
  31. static inline pte_t pte_mkhuge(pte_t pte)
  32. {
  33. /*
  34. * PROT_NONE needs to be remapped from the pte type to the ste type.
  35. * The HW invalid bit is also different for pte and ste. The pte
  36. * invalid bit happens to be the same as the ste _SEGMENT_ENTRY_LARGE
  37. * bit, so we don't have to clear it.
  38. */
  39. if (pte_val(pte) & _PAGE_INVALID) {
  40. if (pte_val(pte) & _PAGE_SWT)
  41. pte_val(pte) |= _HPAGE_TYPE_NONE;
  42. pte_val(pte) |= _SEGMENT_ENTRY_INV;
  43. }
  44. /*
  45. * Clear SW pte bits SWT and SWX, there are no SW bits in a segment
  46. * table entry.
  47. */
  48. pte_val(pte) &= ~(_PAGE_SWT | _PAGE_SWX);
  49. /*
  50. * Also set the change-override bit because we don't need dirty bit
  51. * tracking for hugetlbfs pages.
  52. */
  53. pte_val(pte) |= (_SEGMENT_ENTRY_LARGE | _SEGMENT_ENTRY_CO);
  54. return pte;
  55. }
  56. static inline pte_t huge_pte_wrprotect(pte_t pte)
  57. {
  58. pte_val(pte) |= _PAGE_RO;
  59. return pte;
  60. }
  61. static inline int huge_pte_none(pte_t pte)
  62. {
  63. return (pte_val(pte) & _SEGMENT_ENTRY_INV) &&
  64. !(pte_val(pte) & _SEGMENT_ENTRY_RO);
  65. }
  66. static inline pte_t huge_ptep_get(pte_t *ptep)
  67. {
  68. pte_t pte = *ptep;
  69. unsigned long mask;
  70. if (!MACHINE_HAS_HPAGE) {
  71. ptep = (pte_t *) (pte_val(pte) & _SEGMENT_ENTRY_ORIGIN);
  72. if (ptep) {
  73. mask = pte_val(pte) &
  74. (_SEGMENT_ENTRY_INV | _SEGMENT_ENTRY_RO);
  75. pte = pte_mkhuge(*ptep);
  76. pte_val(pte) |= mask;
  77. }
  78. }
  79. return pte;
  80. }
  81. static inline pte_t huge_ptep_get_and_clear(struct mm_struct *mm,
  82. unsigned long addr, pte_t *ptep)
  83. {
  84. pte_t pte = huge_ptep_get(ptep);
  85. pmd_clear((pmd_t *) ptep);
  86. return pte;
  87. }
  88. static inline void __pmd_csp(pmd_t *pmdp)
  89. {
  90. register unsigned long reg2 asm("2") = pmd_val(*pmdp);
  91. register unsigned long reg3 asm("3") = pmd_val(*pmdp) |
  92. _SEGMENT_ENTRY_INV;
  93. register unsigned long reg4 asm("4") = ((unsigned long) pmdp) + 5;
  94. asm volatile(
  95. " csp %1,%3"
  96. : "=m" (*pmdp)
  97. : "d" (reg2), "d" (reg3), "d" (reg4), "m" (*pmdp) : "cc");
  98. pmd_val(*pmdp) = _SEGMENT_ENTRY_INV | _SEGMENT_ENTRY;
  99. }
  100. static inline void __pmd_idte(unsigned long address, pmd_t *pmdp)
  101. {
  102. unsigned long sto = (unsigned long) pmdp -
  103. pmd_index(address) * sizeof(pmd_t);
  104. if (!(pmd_val(*pmdp) & _SEGMENT_ENTRY_INV)) {
  105. asm volatile(
  106. " .insn rrf,0xb98e0000,%2,%3,0,0"
  107. : "=m" (*pmdp)
  108. : "m" (*pmdp), "a" (sto),
  109. "a" ((address & HPAGE_MASK))
  110. );
  111. }
  112. pmd_val(*pmdp) = _SEGMENT_ENTRY_INV | _SEGMENT_ENTRY;
  113. }
  114. static inline void huge_ptep_invalidate(struct mm_struct *mm,
  115. unsigned long address, pte_t *ptep)
  116. {
  117. pmd_t *pmdp = (pmd_t *) ptep;
  118. if (!MACHINE_HAS_IDTE) {
  119. __pmd_csp(pmdp);
  120. if (mm->context.noexec) {
  121. pmdp = get_shadow_table(pmdp);
  122. __pmd_csp(pmdp);
  123. }
  124. return;
  125. }
  126. __pmd_idte(address, pmdp);
  127. if (mm->context.noexec) {
  128. pmdp = get_shadow_table(pmdp);
  129. __pmd_idte(address, pmdp);
  130. }
  131. return;
  132. }
  133. #define huge_ptep_set_access_flags(__vma, __addr, __ptep, __entry, __dirty) \
  134. ({ \
  135. int __changed = !pte_same(huge_ptep_get(__ptep), __entry); \
  136. if (__changed) { \
  137. huge_ptep_invalidate((__vma)->vm_mm, __addr, __ptep); \
  138. set_huge_pte_at((__vma)->vm_mm, __addr, __ptep, __entry); \
  139. } \
  140. __changed; \
  141. })
  142. #define huge_ptep_set_wrprotect(__mm, __addr, __ptep) \
  143. ({ \
  144. pte_t __pte = huge_ptep_get(__ptep); \
  145. if (pte_write(__pte)) { \
  146. if (atomic_read(&(__mm)->mm_users) > 1 || \
  147. (__mm) != current->active_mm) \
  148. huge_ptep_invalidate(__mm, __addr, __ptep); \
  149. set_huge_pte_at(__mm, __addr, __ptep, \
  150. huge_pte_wrprotect(__pte)); \
  151. } \
  152. })
  153. static inline void huge_ptep_clear_flush(struct vm_area_struct *vma,
  154. unsigned long address, pte_t *ptep)
  155. {
  156. huge_ptep_invalidate(vma->vm_mm, address, ptep);
  157. }
  158. #endif /* _ASM_S390_HUGETLB_H */