|
@@ -6052,7 +6052,7 @@ struct mem_cgroup *parent_mem_cgroup(struct mem_cgroup *memcg)
|
|
|
}
|
|
|
EXPORT_SYMBOL(parent_mem_cgroup);
|
|
|
|
|
|
-static int mem_cgroup_soft_limit_tree_init(void)
|
|
|
+static void __init mem_cgroup_soft_limit_tree_init(void)
|
|
|
{
|
|
|
struct mem_cgroup_tree_per_node *rtpn;
|
|
|
struct mem_cgroup_tree_per_zone *rtpz;
|
|
@@ -6063,8 +6063,7 @@ static int mem_cgroup_soft_limit_tree_init(void)
|
|
|
if (!node_state(node, N_NORMAL_MEMORY))
|
|
|
tmp = -1;
|
|
|
rtpn = kzalloc_node(sizeof(*rtpn), GFP_KERNEL, tmp);
|
|
|
- if (!rtpn)
|
|
|
- goto err_cleanup;
|
|
|
+ BUG_ON(!rtpn);
|
|
|
|
|
|
soft_limit_tree.rb_tree_per_node[node] = rtpn;
|
|
|
|
|
@@ -6074,17 +6073,6 @@ static int mem_cgroup_soft_limit_tree_init(void)
|
|
|
spin_lock_init(&rtpz->lock);
|
|
|
}
|
|
|
}
|
|
|
- return 0;
|
|
|
-
|
|
|
-err_cleanup:
|
|
|
- for_each_node(node) {
|
|
|
- if (!soft_limit_tree.rb_tree_per_node[node])
|
|
|
- break;
|
|
|
- kfree(soft_limit_tree.rb_tree_per_node[node]);
|
|
|
- soft_limit_tree.rb_tree_per_node[node] = NULL;
|
|
|
- }
|
|
|
- return 1;
|
|
|
-
|
|
|
}
|
|
|
|
|
|
static struct cgroup_subsys_state * __ref
|
|
@@ -6106,8 +6094,6 @@ mem_cgroup_css_alloc(struct cgroup *cont)
|
|
|
if (cont->parent == NULL) {
|
|
|
int cpu;
|
|
|
|
|
|
- if (mem_cgroup_soft_limit_tree_init())
|
|
|
- goto free_out;
|
|
|
root_mem_cgroup = memcg;
|
|
|
for_each_possible_cpu(cpu) {
|
|
|
struct memcg_stock_pcp *stock =
|
|
@@ -6850,6 +6836,7 @@ static int __init mem_cgroup_init(void)
|
|
|
{
|
|
|
hotcpu_notifier(memcg_cpu_hotplug_callback, 0);
|
|
|
enable_swap_cgroup();
|
|
|
+ mem_cgroup_soft_limit_tree_init();
|
|
|
return 0;
|
|
|
}
|
|
|
subsys_initcall(mem_cgroup_init);
|