Browse Source

[PATCH] sched: uninline task_rq_lock()

Saves 543 bytes from sched.o (gcc 3.3.3).

Signed-off-by: Oleg Nesterov <oleg@tv-sign.ru>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Cc: Con Kolivas <kernel@kolivas.org>
Cc: Peter Williams <pwil3058@bigpond.net.au>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Oleg Nesterov 19 years ago
parent
commit
9fea80e4d9
1 changed files with 1 additions and 1 deletions
  1. 1 1
      kernel/sched.c

+ 1 - 1
kernel/sched.c

@@ -359,7 +359,7 @@ static inline void finish_lock_switch(runqueue_t *rq, task_t *prev)
  * interrupts.  Note the ordering: we can safely lookup the task_rq without
  * interrupts.  Note the ordering: we can safely lookup the task_rq without
  * explicitly disabling preemption.
  * explicitly disabling preemption.
  */
  */
-static inline runqueue_t *task_rq_lock(task_t *p, unsigned long *flags)
+static runqueue_t *task_rq_lock(task_t *p, unsigned long *flags)
 	__acquires(rq->lock)
 	__acquires(rq->lock)
 {
 {
 	struct runqueue *rq;
 	struct runqueue *rq;