|
@@ -3730,7 +3730,7 @@ int mutex_spin_on_owner(struct mutex *lock, struct thread_info *owner)
|
|
|
* off of preempt_enable. Kernel preemptions off return from interrupt
|
|
|
* occur there and call schedule directly.
|
|
|
*/
|
|
|
-asmlinkage void __sched preempt_schedule(void)
|
|
|
+asmlinkage void __sched notrace preempt_schedule(void)
|
|
|
{
|
|
|
struct thread_info *ti = current_thread_info();
|
|
|
|
|
@@ -3742,9 +3742,9 @@ asmlinkage void __sched preempt_schedule(void)
|
|
|
return;
|
|
|
|
|
|
do {
|
|
|
- add_preempt_count(PREEMPT_ACTIVE);
|
|
|
+ add_preempt_count_notrace(PREEMPT_ACTIVE);
|
|
|
schedule();
|
|
|
- sub_preempt_count(PREEMPT_ACTIVE);
|
|
|
+ sub_preempt_count_notrace(PREEMPT_ACTIVE);
|
|
|
|
|
|
/*
|
|
|
* Check again in case we missed a preemption opportunity
|