hugetlb.h 3.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130
  1. /*
  2. * IBM System z Huge TLB Page Support for Kernel.
  3. *
  4. * Copyright IBM Corp. 2008
  5. * Author(s): Gerald Schaefer <gerald.schaefer@de.ibm.com>
  6. */
  7. #ifndef _ASM_S390_HUGETLB_H
  8. #define _ASM_S390_HUGETLB_H
  9. #include <asm/page.h>
  10. #include <asm/pgtable.h>
  11. #define is_hugepage_only_range(mm, addr, len) 0
  12. #define hugetlb_free_pgd_range free_pgd_range
  13. void set_huge_pte_at(struct mm_struct *mm, unsigned long addr,
  14. pte_t *ptep, pte_t pte);
  15. /*
  16. * If the arch doesn't supply something else, assume that hugepage
  17. * size aligned regions are ok without further preparation.
  18. */
  19. static inline int prepare_hugepage_range(struct file *file,
  20. unsigned long addr, unsigned long len)
  21. {
  22. if (len & ~HPAGE_MASK)
  23. return -EINVAL;
  24. if (addr & ~HPAGE_MASK)
  25. return -EINVAL;
  26. return 0;
  27. }
  28. #define hugetlb_prefault_arch_hook(mm) do { } while (0)
  29. #define arch_clear_hugepage_flags(page) do { } while (0)
  30. int arch_prepare_hugepage(struct page *page);
  31. void arch_release_hugepage(struct page *page);
  32. static inline pte_t huge_pte_wrprotect(pte_t pte)
  33. {
  34. pte_val(pte) |= _PAGE_RO;
  35. return pte;
  36. }
  37. static inline int huge_pte_none(pte_t pte)
  38. {
  39. return (pte_val(pte) & _SEGMENT_ENTRY_INV) &&
  40. !(pte_val(pte) & _SEGMENT_ENTRY_RO);
  41. }
  42. static inline pte_t huge_ptep_get(pte_t *ptep)
  43. {
  44. pte_t pte = *ptep;
  45. unsigned long mask;
  46. if (!MACHINE_HAS_HPAGE) {
  47. ptep = (pte_t *) (pte_val(pte) & _SEGMENT_ENTRY_ORIGIN);
  48. if (ptep) {
  49. mask = pte_val(pte) &
  50. (_SEGMENT_ENTRY_INV | _SEGMENT_ENTRY_RO);
  51. pte = pte_mkhuge(*ptep);
  52. pte_val(pte) |= mask;
  53. }
  54. }
  55. return pte;
  56. }
  57. static inline void __pmd_csp(pmd_t *pmdp)
  58. {
  59. register unsigned long reg2 asm("2") = pmd_val(*pmdp);
  60. register unsigned long reg3 asm("3") = pmd_val(*pmdp) |
  61. _SEGMENT_ENTRY_INV;
  62. register unsigned long reg4 asm("4") = ((unsigned long) pmdp) + 5;
  63. asm volatile(
  64. " csp %1,%3"
  65. : "=m" (*pmdp)
  66. : "d" (reg2), "d" (reg3), "d" (reg4), "m" (*pmdp) : "cc");
  67. }
  68. static inline void huge_ptep_invalidate(struct mm_struct *mm,
  69. unsigned long address, pte_t *ptep)
  70. {
  71. pmd_t *pmdp = (pmd_t *) ptep;
  72. if (MACHINE_HAS_IDTE)
  73. __pmd_idte(address, pmdp);
  74. else
  75. __pmd_csp(pmdp);
  76. pmd_val(*pmdp) = _SEGMENT_ENTRY_INV | _SEGMENT_ENTRY;
  77. }
  78. static inline pte_t huge_ptep_get_and_clear(struct mm_struct *mm,
  79. unsigned long addr, pte_t *ptep)
  80. {
  81. pte_t pte = huge_ptep_get(ptep);
  82. huge_ptep_invalidate(mm, addr, ptep);
  83. return pte;
  84. }
  85. #define huge_ptep_set_access_flags(__vma, __addr, __ptep, __entry, __dirty) \
  86. ({ \
  87. int __changed = !pte_same(huge_ptep_get(__ptep), __entry); \
  88. if (__changed) { \
  89. huge_ptep_invalidate((__vma)->vm_mm, __addr, __ptep); \
  90. set_huge_pte_at((__vma)->vm_mm, __addr, __ptep, __entry); \
  91. } \
  92. __changed; \
  93. })
  94. #define huge_ptep_set_wrprotect(__mm, __addr, __ptep) \
  95. ({ \
  96. pte_t __pte = huge_ptep_get(__ptep); \
  97. if (pte_write(__pte)) { \
  98. huge_ptep_invalidate(__mm, __addr, __ptep); \
  99. set_huge_pte_at(__mm, __addr, __ptep, \
  100. huge_pte_wrprotect(__pte)); \
  101. } \
  102. })
  103. static inline void huge_ptep_clear_flush(struct vm_area_struct *vma,
  104. unsigned long address, pte_t *ptep)
  105. {
  106. huge_ptep_invalidate(vma->vm_mm, address, ptep);
  107. }
  108. #endif /* _ASM_S390_HUGETLB_H */