|
@@ -364,11 +364,10 @@ struct cgroup_subsys_state *mem_cgroup_css(struct mem_cgroup *mem)
|
|
|
}
|
|
|
|
|
|
static struct mem_cgroup_per_zone *
|
|
|
-page_cgroup_zoneinfo(struct page_cgroup *pc)
|
|
|
+page_cgroup_zoneinfo(struct mem_cgroup *mem, struct page *page)
|
|
|
{
|
|
|
- struct mem_cgroup *mem = pc->mem_cgroup;
|
|
|
- int nid = page_cgroup_nid(pc);
|
|
|
- int zid = page_cgroup_zid(pc);
|
|
|
+ int nid = page_to_nid(page);
|
|
|
+ int zid = page_zonenum(page);
|
|
|
|
|
|
return mem_cgroup_zoneinfo(mem, nid, zid);
|
|
|
}
|
|
@@ -800,7 +799,7 @@ void mem_cgroup_del_lru_list(struct page *page, enum lru_list lru)
|
|
|
* We don't check PCG_USED bit. It's cleared when the "page" is finally
|
|
|
* removed from global LRU.
|
|
|
*/
|
|
|
- mz = page_cgroup_zoneinfo(pc);
|
|
|
+ mz = page_cgroup_zoneinfo(pc->mem_cgroup, page);
|
|
|
/* huge page split is done under lru_lock. so, we have no races. */
|
|
|
MEM_CGROUP_ZSTAT(mz, lru) -= 1 << compound_order(page);
|
|
|
if (mem_cgroup_is_root(pc->mem_cgroup))
|
|
@@ -836,7 +835,7 @@ void mem_cgroup_rotate_reclaimable_page(struct page *page)
|
|
|
smp_rmb();
|
|
|
if (mem_cgroup_is_root(pc->mem_cgroup))
|
|
|
return;
|
|
|
- mz = page_cgroup_zoneinfo(pc);
|
|
|
+ mz = page_cgroup_zoneinfo(pc->mem_cgroup, page);
|
|
|
list_move_tail(&pc->lru, &mz->lists[lru]);
|
|
|
}
|
|
|
|
|
@@ -856,7 +855,7 @@ void mem_cgroup_rotate_lru_list(struct page *page, enum lru_list lru)
|
|
|
smp_rmb();
|
|
|
if (mem_cgroup_is_root(pc->mem_cgroup))
|
|
|
return;
|
|
|
- mz = page_cgroup_zoneinfo(pc);
|
|
|
+ mz = page_cgroup_zoneinfo(pc->mem_cgroup, page);
|
|
|
list_move(&pc->lru, &mz->lists[lru]);
|
|
|
}
|
|
|
|
|
@@ -873,7 +872,7 @@ void mem_cgroup_add_lru_list(struct page *page, enum lru_list lru)
|
|
|
return;
|
|
|
/* Ensure pc->mem_cgroup is visible after reading PCG_USED. */
|
|
|
smp_rmb();
|
|
|
- mz = page_cgroup_zoneinfo(pc);
|
|
|
+ mz = page_cgroup_zoneinfo(pc->mem_cgroup, page);
|
|
|
/* huge page split is done under lru_lock. so, we have no races. */
|
|
|
MEM_CGROUP_ZSTAT(mz, lru) += 1 << compound_order(page);
|
|
|
SetPageCgroupAcctLRU(pc);
|
|
@@ -1043,7 +1042,7 @@ mem_cgroup_get_reclaim_stat_from_page(struct page *page)
|
|
|
return NULL;
|
|
|
/* Ensure pc->mem_cgroup is visible after reading PCG_USED. */
|
|
|
smp_rmb();
|
|
|
- mz = page_cgroup_zoneinfo(pc);
|
|
|
+ mz = page_cgroup_zoneinfo(pc->mem_cgroup, page);
|
|
|
if (!mz)
|
|
|
return NULL;
|
|
|
|
|
@@ -2192,7 +2191,7 @@ void mem_cgroup_split_huge_fixup(struct page *head, struct page *tail)
|
|
|
* We hold lru_lock, then, reduce counter directly.
|
|
|
*/
|
|
|
lru = page_lru(head);
|
|
|
- mz = page_cgroup_zoneinfo(head_pc);
|
|
|
+ mz = page_cgroup_zoneinfo(head_pc->mem_cgroup, head);
|
|
|
MEM_CGROUP_ZSTAT(mz, lru) -= 1;
|
|
|
}
|
|
|
tail_pc->flags = head_pc->flags & ~PCGF_NOCOPY_AT_SPLIT;
|