|
@@ -1248,7 +1248,7 @@ int mem_cgroup_prepare_migration(struct page *page, struct mem_cgroup **ptr)
|
|
|
unlock_page_cgroup(pc);
|
|
|
|
|
|
if (mem) {
|
|
|
- ret = mem_cgroup_try_charge(NULL, GFP_HIGHUSER_MOVABLE, &mem);
|
|
|
+ ret = mem_cgroup_try_charge(NULL, GFP_KERNEL, &mem);
|
|
|
css_put(&mem->css);
|
|
|
}
|
|
|
*ptr = mem;
|
|
@@ -1378,7 +1378,7 @@ static int mem_cgroup_resize_limit(struct mem_cgroup *memcg,
|
|
|
break;
|
|
|
|
|
|
progress = try_to_free_mem_cgroup_pages(memcg,
|
|
|
- GFP_HIGHUSER_MOVABLE, false);
|
|
|
+ GFP_KERNEL, false);
|
|
|
if (!progress) retry_count--;
|
|
|
}
|
|
|
return ret;
|
|
@@ -1418,7 +1418,7 @@ int mem_cgroup_resize_memsw_limit(struct mem_cgroup *memcg,
|
|
|
break;
|
|
|
|
|
|
oldusage = res_counter_read_u64(&memcg->memsw, RES_USAGE);
|
|
|
- try_to_free_mem_cgroup_pages(memcg, GFP_HIGHUSER_MOVABLE, true);
|
|
|
+ try_to_free_mem_cgroup_pages(memcg, GFP_KERNEL, true);
|
|
|
curusage = res_counter_read_u64(&memcg->memsw, RES_USAGE);
|
|
|
if (curusage >= oldusage)
|
|
|
retry_count--;
|
|
@@ -1464,7 +1464,7 @@ static int mem_cgroup_force_empty_list(struct mem_cgroup *mem,
|
|
|
}
|
|
|
spin_unlock_irqrestore(&zone->lru_lock, flags);
|
|
|
|
|
|
- ret = mem_cgroup_move_parent(pc, mem, GFP_HIGHUSER_MOVABLE);
|
|
|
+ ret = mem_cgroup_move_parent(pc, mem, GFP_KERNEL);
|
|
|
if (ret == -ENOMEM)
|
|
|
break;
|
|
|
|
|
@@ -1550,7 +1550,7 @@ try_to_free:
|
|
|
goto out;
|
|
|
}
|
|
|
progress = try_to_free_mem_cgroup_pages(mem,
|
|
|
- GFP_HIGHUSER_MOVABLE, false);
|
|
|
+ GFP_KERNEL, false);
|
|
|
if (!progress) {
|
|
|
nr_retries--;
|
|
|
/* maybe some writeback is necessary */
|