|
@@ -738,7 +738,7 @@ static int __do_huge_pmd_anonymous_page(struct mm_struct *mm,
|
|
|
pgtable_trans_huge_deposit(mm, pmd, pgtable);
|
|
|
set_pmd_at(mm, haddr, pmd, entry);
|
|
|
add_mm_counter(mm, MM_ANONPAGES, HPAGE_PMD_NR);
|
|
|
- mm->nr_ptes++;
|
|
|
+ atomic_long_inc(&mm->nr_ptes);
|
|
|
spin_unlock(&mm->page_table_lock);
|
|
|
}
|
|
|
|
|
@@ -771,7 +771,7 @@ static bool set_huge_zero_page(pgtable_t pgtable, struct mm_struct *mm,
|
|
|
entry = pmd_mkhuge(entry);
|
|
|
pgtable_trans_huge_deposit(mm, pmd, pgtable);
|
|
|
set_pmd_at(mm, haddr, pmd, entry);
|
|
|
- mm->nr_ptes++;
|
|
|
+ atomic_long_inc(&mm->nr_ptes);
|
|
|
return true;
|
|
|
}
|
|
|
|
|
@@ -896,7 +896,7 @@ int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm,
|
|
|
pmd = pmd_mkold(pmd_wrprotect(pmd));
|
|
|
pgtable_trans_huge_deposit(dst_mm, dst_pmd, pgtable);
|
|
|
set_pmd_at(dst_mm, addr, dst_pmd, pmd);
|
|
|
- dst_mm->nr_ptes++;
|
|
|
+ atomic_long_inc(&dst_mm->nr_ptes);
|
|
|
|
|
|
ret = 0;
|
|
|
out_unlock:
|
|
@@ -1392,7 +1392,7 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma,
|
|
|
tlb_remove_pmd_tlb_entry(tlb, pmd, addr);
|
|
|
pgtable = pgtable_trans_huge_withdraw(tlb->mm, pmd);
|
|
|
if (is_huge_zero_pmd(orig_pmd)) {
|
|
|
- tlb->mm->nr_ptes--;
|
|
|
+ atomic_long_dec(&tlb->mm->nr_ptes);
|
|
|
spin_unlock(&tlb->mm->page_table_lock);
|
|
|
put_huge_zero_page();
|
|
|
} else {
|
|
@@ -1401,7 +1401,7 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma,
|
|
|
VM_BUG_ON(page_mapcount(page) < 0);
|
|
|
add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR);
|
|
|
VM_BUG_ON(!PageHead(page));
|
|
|
- tlb->mm->nr_ptes--;
|
|
|
+ atomic_long_dec(&tlb->mm->nr_ptes);
|
|
|
spin_unlock(&tlb->mm->page_table_lock);
|
|
|
tlb_remove_page(tlb, page);
|
|
|
}
|