|
@@ -983,7 +983,7 @@ void set_task_cpu(struct task_struct *p, unsigned int new_cpu)
|
|
|
* ttwu() will sort out the placement.
|
|
|
*/
|
|
|
WARN_ON_ONCE(p->state != TASK_RUNNING && p->state != TASK_WAKING &&
|
|
|
- !(task_thread_info(p)->preempt_count & PREEMPT_ACTIVE));
|
|
|
+ !(task_preempt_count(p) & PREEMPT_ACTIVE));
|
|
|
|
|
|
#ifdef CONFIG_LOCKDEP
|
|
|
/*
|
|
@@ -1723,8 +1723,7 @@ void sched_fork(struct task_struct *p)
|
|
|
p->on_cpu = 0;
|
|
|
#endif
|
|
|
#ifdef CONFIG_PREEMPT_COUNT
|
|
|
- /* Want to start with kernel preemption disabled. */
|
|
|
- task_thread_info(p)->preempt_count = PREEMPT_DISABLED;
|
|
|
+ init_task_preempt_count(p);
|
|
|
#endif
|
|
|
#ifdef CONFIG_SMP
|
|
|
plist_node_init(&p->pushable_tasks, MAX_PRIO);
|
|
@@ -4217,7 +4216,7 @@ void init_idle(struct task_struct *idle, int cpu)
|
|
|
raw_spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
|
|
/* Set the preempt count _outside_ the spinlocks! */
|
|
|
- task_thread_info(idle)->preempt_count = PREEMPT_ENABLED;
|
|
|
+ init_idle_preempt_count(idle, cpu);
|
|
|
|
|
|
/*
|
|
|
* The idle tasks have their own, simple scheduling class:
|