|
@@ -3519,12 +3519,12 @@ static int do_nonlinear_fault(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
}
|
|
|
|
|
|
int numa_migrate_prep(struct page *page, struct vm_area_struct *vma,
|
|
|
- unsigned long addr, int current_nid)
|
|
|
+ unsigned long addr, int page_nid)
|
|
|
{
|
|
|
get_page(page);
|
|
|
|
|
|
count_vm_numa_event(NUMA_HINT_FAULTS);
|
|
|
- if (current_nid == numa_node_id())
|
|
|
+ if (page_nid == numa_node_id())
|
|
|
count_vm_numa_event(NUMA_HINT_FAULTS_LOCAL);
|
|
|
|
|
|
return mpol_misplaced(page, vma, addr);
|
|
@@ -3535,7 +3535,7 @@ int do_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
{
|
|
|
struct page *page = NULL;
|
|
|
spinlock_t *ptl;
|
|
|
- int current_nid = -1;
|
|
|
+ int page_nid = -1;
|
|
|
int target_nid;
|
|
|
bool migrated = false;
|
|
|
|
|
@@ -3565,15 +3565,10 @@ int do_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
- current_nid = page_to_nid(page);
|
|
|
- target_nid = numa_migrate_prep(page, vma, addr, current_nid);
|
|
|
+ page_nid = page_to_nid(page);
|
|
|
+ target_nid = numa_migrate_prep(page, vma, addr, page_nid);
|
|
|
pte_unmap_unlock(ptep, ptl);
|
|
|
if (target_nid == -1) {
|
|
|
- /*
|
|
|
- * Account for the fault against the current node if it not
|
|
|
- * being replaced regardless of where the page is located.
|
|
|
- */
|
|
|
- current_nid = numa_node_id();
|
|
|
put_page(page);
|
|
|
goto out;
|
|
|
}
|
|
@@ -3581,11 +3576,11 @@ int do_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
/* Migrate to the requested node */
|
|
|
migrated = migrate_misplaced_page(page, target_nid);
|
|
|
if (migrated)
|
|
|
- current_nid = target_nid;
|
|
|
+ page_nid = target_nid;
|
|
|
|
|
|
out:
|
|
|
- if (current_nid != -1)
|
|
|
- task_numa_fault(current_nid, 1, migrated);
|
|
|
+ if (page_nid != -1)
|
|
|
+ task_numa_fault(page_nid, 1, migrated);
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
@@ -3600,7 +3595,6 @@ static int do_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
unsigned long offset;
|
|
|
spinlock_t *ptl;
|
|
|
bool numa = false;
|
|
|
- int local_nid = numa_node_id();
|
|
|
|
|
|
spin_lock(&mm->page_table_lock);
|
|
|
pmd = *pmdp;
|
|
@@ -3623,9 +3617,10 @@ static int do_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
for (addr = _addr + offset; addr < _addr + PMD_SIZE; pte++, addr += PAGE_SIZE) {
|
|
|
pte_t pteval = *pte;
|
|
|
struct page *page;
|
|
|
- int curr_nid = local_nid;
|
|
|
+ int page_nid = -1;
|
|
|
int target_nid;
|
|
|
- bool migrated;
|
|
|
+ bool migrated = false;
|
|
|
+
|
|
|
if (!pte_present(pteval))
|
|
|
continue;
|
|
|
if (!pte_numa(pteval))
|
|
@@ -3647,25 +3642,19 @@ static int do_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
|
|
|
if (unlikely(page_mapcount(page) != 1))
|
|
|
continue;
|
|
|
|
|
|
- /*
|
|
|
- * Note that the NUMA fault is later accounted to either
|
|
|
- * the node that is currently running or where the page is
|
|
|
- * migrated to.
|
|
|
- */
|
|
|
- curr_nid = local_nid;
|
|
|
- target_nid = numa_migrate_prep(page, vma, addr,
|
|
|
- page_to_nid(page));
|
|
|
- if (target_nid == -1) {
|
|
|
+ page_nid = page_to_nid(page);
|
|
|
+ target_nid = numa_migrate_prep(page, vma, addr, page_nid);
|
|
|
+ pte_unmap_unlock(pte, ptl);
|
|
|
+ if (target_nid != -1) {
|
|
|
+ migrated = migrate_misplaced_page(page, target_nid);
|
|
|
+ if (migrated)
|
|
|
+ page_nid = target_nid;
|
|
|
+ } else {
|
|
|
put_page(page);
|
|
|
- continue;
|
|
|
}
|
|
|
|
|
|
- /* Migrate to the requested node */
|
|
|
- pte_unmap_unlock(pte, ptl);
|
|
|
- migrated = migrate_misplaced_page(page, target_nid);
|
|
|
- if (migrated)
|
|
|
- curr_nid = target_nid;
|
|
|
- task_numa_fault(curr_nid, 1, migrated);
|
|
|
+ if (page_nid != -1)
|
|
|
+ task_numa_fault(page_nid, 1, migrated);
|
|
|
|
|
|
pte = pte_offset_map_lock(mm, pmdp, addr, &ptl);
|
|
|
}
|