|
@@ -1969,11 +1969,10 @@ context_switch(struct rq *rq, struct task_struct *prev,
|
|
|
}
|
|
|
|
|
|
/*
|
|
|
- * nr_running, nr_uninterruptible and nr_context_switches:
|
|
|
+ * nr_running and nr_context_switches:
|
|
|
*
|
|
|
* externally visible scheduler statistics: current number of runnable
|
|
|
- * threads, current number of uninterruptible-sleeping threads, total
|
|
|
- * number of context switches performed since bootup.
|
|
|
+ * threads, total number of context switches performed since bootup.
|
|
|
*/
|
|
|
unsigned long nr_running(void)
|
|
|
{
|
|
@@ -1985,23 +1984,6 @@ unsigned long nr_running(void)
|
|
|
return sum;
|
|
|
}
|
|
|
|
|
|
-unsigned long nr_uninterruptible(void)
|
|
|
-{
|
|
|
- unsigned long i, sum = 0;
|
|
|
-
|
|
|
- for_each_possible_cpu(i)
|
|
|
- sum += cpu_rq(i)->nr_uninterruptible;
|
|
|
-
|
|
|
- /*
|
|
|
- * Since we read the counters lockless, it might be slightly
|
|
|
- * inaccurate. Do not allow it to go below zero though:
|
|
|
- */
|
|
|
- if (unlikely((long)sum < 0))
|
|
|
- sum = 0;
|
|
|
-
|
|
|
- return sum;
|
|
|
-}
|
|
|
-
|
|
|
unsigned long long nr_context_switches(void)
|
|
|
{
|
|
|
int i;
|