|
@@ -170,8 +170,13 @@ static inline void _local_bh_enable_ip(unsigned long ip)
|
|
|
*/
|
|
|
sub_preempt_count(SOFTIRQ_DISABLE_OFFSET - 1);
|
|
|
|
|
|
- if (unlikely(!in_interrupt() && local_softirq_pending()))
|
|
|
+ if (unlikely(!in_interrupt() && local_softirq_pending())) {
|
|
|
+ /*
|
|
|
+ * Run softirq if any pending. And do it in its own stack
|
|
|
+ * as we may be calling this deep in a task call stack already.
|
|
|
+ */
|
|
|
do_softirq();
|
|
|
+ }
|
|
|
|
|
|
dec_preempt_count();
|
|
|
#ifdef CONFIG_TRACE_IRQFLAGS
|
|
@@ -769,6 +774,10 @@ static void run_ksoftirqd(unsigned int cpu)
|
|
|
{
|
|
|
local_irq_disable();
|
|
|
if (local_softirq_pending()) {
|
|
|
+ /*
|
|
|
+ * We can safely run softirq on inline stack, as we are not deep
|
|
|
+ * in the task stack here.
|
|
|
+ */
|
|
|
__do_softirq();
|
|
|
rcu_note_context_switch(cpu);
|
|
|
local_irq_enable();
|