|
@@ -4121,12 +4121,6 @@ static inline void schedule_debug(struct task_struct *prev)
|
|
|
profile_hit(SCHED_PROFILING, __builtin_return_address(0));
|
|
|
|
|
|
schedstat_inc(this_rq(), sched_count);
|
|
|
-#ifdef CONFIG_SCHEDSTATS
|
|
|
- if (unlikely(prev->lock_depth >= 0)) {
|
|
|
- schedstat_inc(this_rq(), rq_sched_info.bkl_count);
|
|
|
- schedstat_inc(prev, sched_info.bkl_count);
|
|
|
- }
|
|
|
-#endif
|
|
|
}
|
|
|
|
|
|
static void put_prev_task(struct rq *rq, struct task_struct *prev)
|
|
@@ -5852,11 +5846,8 @@ void __cpuinit init_idle(struct task_struct *idle, int cpu)
|
|
|
raw_spin_unlock_irqrestore(&rq->lock, flags);
|
|
|
|
|
|
/* Set the preempt count _outside_ the spinlocks! */
|
|
|
-#if defined(CONFIG_PREEMPT)
|
|
|
- task_thread_info(idle)->preempt_count = (idle->lock_depth >= 0);
|
|
|
-#else
|
|
|
task_thread_info(idle)->preempt_count = 0;
|
|
|
-#endif
|
|
|
+
|
|
|
/*
|
|
|
* The idle tasks have their own, simple scheduling class:
|
|
|
*/
|