|
@@ -2116,7 +2116,7 @@ int cgroup_attach_proc(struct cgroup *cgrp, struct task_struct *leader)
|
|
|
retval = -EAGAIN;
|
|
|
goto out_free_group_list;
|
|
|
}
|
|
|
- /* take a reference on each task in the group to go in the array. */
|
|
|
+
|
|
|
tsk = leader;
|
|
|
i = nr_migrating_tasks = 0;
|
|
|
do {
|
|
@@ -2128,7 +2128,6 @@ int cgroup_attach_proc(struct cgroup *cgrp, struct task_struct *leader)
|
|
|
|
|
|
/* as per above, nr_threads may decrease, but not increase. */
|
|
|
BUG_ON(i >= group_size);
|
|
|
- get_task_struct(tsk);
|
|
|
/*
|
|
|
* saying GFP_ATOMIC has no effect here because we did prealloc
|
|
|
* earlier, but it's good form to communicate our expectations.
|
|
@@ -2150,7 +2149,7 @@ int cgroup_attach_proc(struct cgroup *cgrp, struct task_struct *leader)
|
|
|
/* methods shouldn't be called if no task is actually migrating */
|
|
|
retval = 0;
|
|
|
if (!nr_migrating_tasks)
|
|
|
- goto out_put_tasks;
|
|
|
+ goto out_free_group_list;
|
|
|
|
|
|
/*
|
|
|
* step 1: check that we can legitimately attach to the cgroup.
|
|
@@ -2234,12 +2233,6 @@ out_cancel_attach:
|
|
|
ss->cancel_attach(ss, cgrp, &tset);
|
|
|
}
|
|
|
}
|
|
|
-out_put_tasks:
|
|
|
- /* clean up the array of referenced threads in the group. */
|
|
|
- for (i = 0; i < group_size; i++) {
|
|
|
- tc = flex_array_get(group, i);
|
|
|
- put_task_struct(tc->task);
|
|
|
- }
|
|
|
out_free_group_list:
|
|
|
flex_array_free(group);
|
|
|
return retval;
|