|
@@ -723,9 +723,6 @@ static void dequeue_task(struct rq *rq, struct task_struct *p, int flags)
|
|
|
p->sched_class->dequeue_task(rq, p, flags);
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * activate_task - move a task to the runqueue.
|
|
|
- */
|
|
|
void activate_task(struct rq *rq, struct task_struct *p, int flags)
|
|
|
{
|
|
|
if (task_contributes_to_load(p))
|
|
@@ -734,9 +731,6 @@ void activate_task(struct rq *rq, struct task_struct *p, int flags)
|
|
|
enqueue_task(rq, p, flags);
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * deactivate_task - remove a task from the runqueue.
|
|
|
- */
|
|
|
void deactivate_task(struct rq *rq, struct task_struct *p, int flags)
|
|
|
{
|
|
|
if (task_contributes_to_load(p))
|
|
@@ -4134,7 +4128,7 @@ recheck:
|
|
|
on_rq = p->on_rq;
|
|
|
running = task_current(rq, p);
|
|
|
if (on_rq)
|
|
|
- deactivate_task(rq, p, 0);
|
|
|
+ dequeue_task(rq, p, 0);
|
|
|
if (running)
|
|
|
p->sched_class->put_prev_task(rq, p);
|
|
|
|
|
@@ -4147,7 +4141,7 @@ recheck:
|
|
|
if (running)
|
|
|
p->sched_class->set_curr_task(rq);
|
|
|
if (on_rq)
|
|
|
- activate_task(rq, p, 0);
|
|
|
+ enqueue_task(rq, p, 0);
|
|
|
|
|
|
check_class_changed(rq, p, prev_class, oldprio);
|
|
|
task_rq_unlock(rq, p, &flags);
|
|
@@ -4998,9 +4992,9 @@ static int __migrate_task(struct task_struct *p, int src_cpu, int dest_cpu)
|
|
|
* placed properly.
|
|
|
*/
|
|
|
if (p->on_rq) {
|
|
|
- deactivate_task(rq_src, p, 0);
|
|
|
+ dequeue_task(rq_src, p, 0);
|
|
|
set_task_cpu(p, dest_cpu);
|
|
|
- activate_task(rq_dest, p, 0);
|
|
|
+ enqueue_task(rq_dest, p, 0);
|
|
|
check_preempt_curr(rq_dest, p, 0);
|
|
|
}
|
|
|
done:
|
|
@@ -7032,10 +7026,10 @@ static void normalize_task(struct rq *rq, struct task_struct *p)
|
|
|
|
|
|
on_rq = p->on_rq;
|
|
|
if (on_rq)
|
|
|
- deactivate_task(rq, p, 0);
|
|
|
+ dequeue_task(rq, p, 0);
|
|
|
__setscheduler(rq, p, SCHED_NORMAL, 0);
|
|
|
if (on_rq) {
|
|
|
- activate_task(rq, p, 0);
|
|
|
+ enqueue_task(rq, p, 0);
|
|
|
resched_task(rq->curr);
|
|
|
}
|
|
|
|