|
@@ -6335,6 +6335,30 @@ static void calc_global_load_remove(struct rq *rq)
|
|
|
rq->calc_load_active = 0;
|
|
|
}
|
|
|
|
|
|
+#ifdef CONFIG_CFS_BANDWIDTH
|
|
|
+static void unthrottle_offline_cfs_rqs(struct rq *rq)
|
|
|
+{
|
|
|
+ struct cfs_rq *cfs_rq;
|
|
|
+
|
|
|
+ for_each_leaf_cfs_rq(rq, cfs_rq) {
|
|
|
+ struct cfs_bandwidth *cfs_b = tg_cfs_bandwidth(cfs_rq->tg);
|
|
|
+
|
|
|
+ if (!cfs_rq->runtime_enabled)
|
|
|
+ continue;
|
|
|
+
|
|
|
+ /*
|
|
|
+ * clock_task is not advancing so we just need to make sure
|
|
|
+ * there's some valid quota amount
|
|
|
+ */
|
|
|
+ cfs_rq->runtime_remaining = cfs_b->quota;
|
|
|
+ if (cfs_rq_throttled(cfs_rq))
|
|
|
+ unthrottle_cfs_rq(cfs_rq);
|
|
|
+ }
|
|
|
+}
|
|
|
+#else
|
|
|
+static void unthrottle_offline_cfs_rqs(struct rq *rq) {}
|
|
|
+#endif
|
|
|
+
|
|
|
/*
|
|
|
* Migrate all tasks from the rq, sleeping tasks will be migrated by
|
|
|
* try_to_wake_up()->select_task_rq().
|
|
@@ -6360,6 +6384,9 @@ static void migrate_tasks(unsigned int dead_cpu)
|
|
|
*/
|
|
|
rq->stop = NULL;
|
|
|
|
|
|
+ /* Ensure any throttled groups are reachable by pick_next_task */
|
|
|
+ unthrottle_offline_cfs_rqs(rq);
|
|
|
+
|
|
|
for ( ; ; ) {
|
|
|
/*
|
|
|
* There's this thread running, bail when that's the only
|