|
@@ -538,22 +538,19 @@ extern int group_balance_cpu(struct sched_group *sg);
|
|
|
/*
|
|
|
* Return the group to which this tasks belongs.
|
|
|
*
|
|
|
- * We use task_subsys_state_check() and extend the RCU verification with
|
|
|
- * pi->lock and rq->lock because cpu_cgroup_attach() holds those locks for each
|
|
|
- * task it moves into the cgroup. Therefore by holding either of those locks,
|
|
|
- * we pin the task to the current cgroup.
|
|
|
+ * We cannot use task_subsys_state() and friends because the cgroup
|
|
|
+ * subsystem changes that value before the cgroup_subsys::attach() method
|
|
|
+ * is called, therefore we cannot pin it and might observe the wrong value.
|
|
|
+ *
|
|
|
+ * The same is true for autogroup's p->signal->autogroup->tg, the autogroup
|
|
|
+ * core changes this before calling sched_move_task().
|
|
|
+ *
|
|
|
+ * Instead we use a 'copy' which is updated from sched_move_task() while
|
|
|
+ * holding both task_struct::pi_lock and rq::lock.
|
|
|
*/
|
|
|
static inline struct task_group *task_group(struct task_struct *p)
|
|
|
{
|
|
|
- struct task_group *tg;
|
|
|
- struct cgroup_subsys_state *css;
|
|
|
-
|
|
|
- css = task_subsys_state_check(p, cpu_cgroup_subsys_id,
|
|
|
- lockdep_is_held(&p->pi_lock) ||
|
|
|
- lockdep_is_held(&task_rq(p)->lock));
|
|
|
- tg = container_of(css, struct task_group, css);
|
|
|
-
|
|
|
- return autogroup_task_group(p, tg);
|
|
|
+ return p->sched_task_group;
|
|
|
}
|
|
|
|
|
|
/* Change a task's cfs_rq and parent entity if it moves across CPUs/groups */
|