|
@@ -2425,14 +2425,14 @@ static inline void update_sg_lb_stats(struct sched_domain *sd,
|
|
|
* domains. In the newly idle case, we will allow all the cpu's
|
|
|
* to do the newly idle load balance.
|
|
|
*/
|
|
|
- if (idle != CPU_NEWLY_IDLE && local_group &&
|
|
|
- balance_cpu != this_cpu) {
|
|
|
- *balance = 0;
|
|
|
- return;
|
|
|
+ if (idle != CPU_NEWLY_IDLE && local_group) {
|
|
|
+ if (balance_cpu != this_cpu) {
|
|
|
+ *balance = 0;
|
|
|
+ return;
|
|
|
+ }
|
|
|
+ update_group_power(sd, this_cpu);
|
|
|
}
|
|
|
|
|
|
- update_group_power(sd, this_cpu);
|
|
|
-
|
|
|
/* Adjust by relative CPU power of the group */
|
|
|
sgs->avg_load = (sgs->group_load * SCHED_LOAD_SCALE) / group->cpu_power;
|
|
|
|