Sfoglia il codice sorgente

block: let io_schedule() flush the plug inline

Linus correctly observes that the most important dispatch cases
are now done from kblockd, this isn't ideal for latency reasons.
The original reason for switching dispatches out-of-line was to
avoid too deep a stack, so by _only_ letting the "accidental"
flush directly in schedule() be guarded by offload to kblockd,
we should be able to get the best of both worlds.

So add a blk_schedule_flush_plug() that offloads to kblockd,
and only use that from the schedule() path.

Signed-off-by: Jens Axboe <jaxboe@fusionio.com>
Jens Axboe 14 anni fa
parent
commit
a237c1c5bc
2 ha cambiato i file con 14 aggiunte e 1 eliminazioni
  1. 13 0
      include/linux/blkdev.h
  2. 1 1
      kernel/sched.c

+ 13 - 0
include/linux/blkdev.h

@@ -871,6 +871,14 @@ static inline void blk_flush_plug(struct task_struct *tsk)
 {
 {
 	struct blk_plug *plug = tsk->plug;
 	struct blk_plug *plug = tsk->plug;
 
 
+	if (plug)
+		blk_flush_plug_list(plug, false);
+}
+
+static inline void blk_schedule_flush_plug(struct task_struct *tsk)
+{
+	struct blk_plug *plug = tsk->plug;
+
 	if (plug)
 	if (plug)
 		blk_flush_plug_list(plug, true);
 		blk_flush_plug_list(plug, true);
 }
 }
@@ -1317,6 +1325,11 @@ static inline void blk_flush_plug(struct task_struct *task)
 {
 {
 }
 }
 
 
+static inline void blk_schedule_flush_plug(struct task_struct *task)
+{
+}
+
+
 static inline bool blk_needs_flush_plug(struct task_struct *tsk)
 static inline bool blk_needs_flush_plug(struct task_struct *tsk)
 {
 {
 	return false;
 	return false;

+ 1 - 1
kernel/sched.c

@@ -4118,7 +4118,7 @@ need_resched:
 			 */
 			 */
 			if (blk_needs_flush_plug(prev)) {
 			if (blk_needs_flush_plug(prev)) {
 				raw_spin_unlock(&rq->lock);
 				raw_spin_unlock(&rq->lock);
-				blk_flush_plug(prev);
+				blk_schedule_flush_plug(prev);
 				raw_spin_lock(&rq->lock);
 				raw_spin_lock(&rq->lock);
 			}
 			}
 		}
 		}