|
@@ -857,8 +857,24 @@ static int sched_rt_runtime_exceeded(struct rt_rq *rt_rq)
|
|
|
return 0;
|
|
|
|
|
|
if (rt_rq->rt_time > runtime) {
|
|
|
- rt_rq->rt_throttled = 1;
|
|
|
- printk_once(KERN_WARNING "sched: RT throttling activated\n");
|
|
|
+ struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq);
|
|
|
+
|
|
|
+ /*
|
|
|
+ * Don't actually throttle groups that have no runtime assigned
|
|
|
+ * but accrue some time due to boosting.
|
|
|
+ */
|
|
|
+ if (likely(rt_b->rt_runtime)) {
|
|
|
+ rt_rq->rt_throttled = 1;
|
|
|
+ printk_once(KERN_WARNING "sched: RT throttling activated\n");
|
|
|
+ } else {
|
|
|
+ /*
|
|
|
+ * In case we did anyway, make it go away,
|
|
|
+ * replenishment is a joke, since it will replenish us
|
|
|
+ * with exactly 0 ns.
|
|
|
+ */
|
|
|
+ rt_rq->rt_time = 0;
|
|
|
+ }
|
|
|
+
|
|
|
if (rt_rq_throttled(rt_rq)) {
|
|
|
sched_rt_rq_dequeue(rt_rq);
|
|
|
return 1;
|