|
@@ -88,6 +88,15 @@ static struct dbs_tuners dbs_tuners_ins = {
|
|
|
.sampling_down_factor = DEF_SAMPLING_DOWN_FACTOR,
|
|
|
};
|
|
|
|
|
|
+static inline unsigned int get_cpu_idle_time(unsigned int cpu)
|
|
|
+{
|
|
|
+ return kstat_cpu(cpu).cpustat.idle +
|
|
|
+ kstat_cpu(cpu).cpustat.iowait +
|
|
|
+ ( !dbs_tuners_ins.ignore_nice ?
|
|
|
+ kstat_cpu(cpu).cpustat.nice :
|
|
|
+ 0);
|
|
|
+}
|
|
|
+
|
|
|
/************************** sysfs interface ************************/
|
|
|
static ssize_t show_sampling_rate_max(struct cpufreq_policy *policy, char *buf)
|
|
|
{
|
|
@@ -220,16 +229,10 @@ static ssize_t store_ignore_nice(struct cpufreq_policy *policy,
|
|
|
dbs_tuners_ins.ignore_nice = input;
|
|
|
|
|
|
/* we need to re-evaluate prev_cpu_idle_up and prev_cpu_idle_down */
|
|
|
- for_each_cpu_mask(j, policy->cpus) {
|
|
|
+ for_each_online_cpu(j) {
|
|
|
struct cpu_dbs_info_s *j_dbs_info;
|
|
|
j_dbs_info = &per_cpu(cpu_dbs_info, j);
|
|
|
- j_dbs_info->cur_policy = policy;
|
|
|
-
|
|
|
- j_dbs_info->prev_cpu_idle_up =
|
|
|
- kstat_cpu(j).cpustat.idle +
|
|
|
- kstat_cpu(j).cpustat.iowait +
|
|
|
- ( !dbs_tuners_ins.ignore_nice
|
|
|
- ? kstat_cpu(j).cpustat.nice : 0 );
|
|
|
+ j_dbs_info->prev_cpu_idle_up = get_cpu_idle_time(j);
|
|
|
j_dbs_info->prev_cpu_idle_down = j_dbs_info->prev_cpu_idle_up;
|
|
|
}
|
|
|
up(&dbs_sem);
|
|
@@ -322,15 +325,10 @@ static void dbs_check_cpu(int cpu)
|
|
|
*/
|
|
|
|
|
|
/* Check for frequency increase */
|
|
|
- total_idle_ticks = kstat_cpu(cpu).cpustat.idle +
|
|
|
- kstat_cpu(cpu).cpustat.iowait;
|
|
|
- /* consider 'nice' tasks as 'idle' time too if required */
|
|
|
- if (dbs_tuners_ins.ignore_nice == 0)
|
|
|
- total_idle_ticks += kstat_cpu(cpu).cpustat.nice;
|
|
|
+ total_idle_ticks = get_cpu_idle_time(cpu);
|
|
|
idle_ticks = total_idle_ticks -
|
|
|
this_dbs_info->prev_cpu_idle_up;
|
|
|
this_dbs_info->prev_cpu_idle_up = total_idle_ticks;
|
|
|
-
|
|
|
|
|
|
for_each_cpu_mask(j, policy->cpus) {
|
|
|
unsigned int tmp_idle_ticks;
|
|
@@ -341,11 +339,7 @@ static void dbs_check_cpu(int cpu)
|
|
|
|
|
|
j_dbs_info = &per_cpu(cpu_dbs_info, j);
|
|
|
/* Check for frequency increase */
|
|
|
- total_idle_ticks = kstat_cpu(j).cpustat.idle +
|
|
|
- kstat_cpu(j).cpustat.iowait;
|
|
|
- /* consider 'nice' too? */
|
|
|
- if (dbs_tuners_ins.ignore_nice == 0)
|
|
|
- total_idle_ticks += kstat_cpu(j).cpustat.nice;
|
|
|
+ total_idle_ticks = get_cpu_idle_time(j);
|
|
|
tmp_idle_ticks = total_idle_ticks -
|
|
|
j_dbs_info->prev_cpu_idle_up;
|
|
|
j_dbs_info->prev_cpu_idle_up = total_idle_ticks;
|
|
@@ -360,14 +354,14 @@ static void dbs_check_cpu(int cpu)
|
|
|
usecs_to_jiffies(dbs_tuners_ins.sampling_rate);
|
|
|
|
|
|
if (idle_ticks < up_idle_ticks) {
|
|
|
+ down_skip[cpu] = 0;
|
|
|
+ this_dbs_info->prev_cpu_idle_down = total_idle_ticks;
|
|
|
/* if we are already at full speed then break out early */
|
|
|
if (policy->cur == policy->max)
|
|
|
return;
|
|
|
|
|
|
__cpufreq_driver_target(policy, policy->max,
|
|
|
CPUFREQ_RELATION_H);
|
|
|
- down_skip[cpu] = 0;
|
|
|
- this_dbs_info->prev_cpu_idle_down = total_idle_ticks;
|
|
|
return;
|
|
|
}
|
|
|
|
|
@@ -376,11 +370,7 @@ static void dbs_check_cpu(int cpu)
|
|
|
if (down_skip[cpu] < dbs_tuners_ins.sampling_down_factor)
|
|
|
return;
|
|
|
|
|
|
- total_idle_ticks = kstat_cpu(cpu).cpustat.idle +
|
|
|
- kstat_cpu(cpu).cpustat.iowait;
|
|
|
- /* consider 'nice' too? */
|
|
|
- if (dbs_tuners_ins.ignore_nice == 0)
|
|
|
- total_idle_ticks += kstat_cpu(cpu).cpustat.nice;
|
|
|
+ total_idle_ticks = this_dbs_info->prev_cpu_idle_up;
|
|
|
idle_ticks = total_idle_ticks -
|
|
|
this_dbs_info->prev_cpu_idle_down;
|
|
|
this_dbs_info->prev_cpu_idle_down = total_idle_ticks;
|
|
@@ -393,12 +383,8 @@ static void dbs_check_cpu(int cpu)
|
|
|
continue;
|
|
|
|
|
|
j_dbs_info = &per_cpu(cpu_dbs_info, j);
|
|
|
- /* Check for frequency increase */
|
|
|
- total_idle_ticks = kstat_cpu(j).cpustat.idle +
|
|
|
- kstat_cpu(j).cpustat.iowait;
|
|
|
- /* consider 'nice' too? */
|
|
|
- if (dbs_tuners_ins.ignore_nice == 0)
|
|
|
- total_idle_ticks += kstat_cpu(j).cpustat.nice;
|
|
|
+ /* Check for frequency decrease */
|
|
|
+ total_idle_ticks = j_dbs_info->prev_cpu_idle_up;
|
|
|
tmp_idle_ticks = total_idle_ticks -
|
|
|
j_dbs_info->prev_cpu_idle_down;
|
|
|
j_dbs_info->prev_cpu_idle_down = total_idle_ticks;
|
|
@@ -414,7 +400,7 @@ static void dbs_check_cpu(int cpu)
|
|
|
freq_down_sampling_rate = dbs_tuners_ins.sampling_rate *
|
|
|
dbs_tuners_ins.sampling_down_factor;
|
|
|
down_idle_ticks = (100 - dbs_tuners_ins.down_threshold) *
|
|
|
- usecs_to_jiffies(freq_down_sampling_rate);
|
|
|
+ usecs_to_jiffies(freq_down_sampling_rate);
|
|
|
|
|
|
if (idle_ticks > down_idle_ticks ) {
|
|
|
/* if we are already at the lowest speed then break out early
|
|
@@ -488,11 +474,7 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy,
|
|
|
j_dbs_info = &per_cpu(cpu_dbs_info, j);
|
|
|
j_dbs_info->cur_policy = policy;
|
|
|
|
|
|
- j_dbs_info->prev_cpu_idle_up =
|
|
|
- kstat_cpu(j).cpustat.idle +
|
|
|
- kstat_cpu(j).cpustat.iowait +
|
|
|
- ( !dbs_tuners_ins.ignore_nice
|
|
|
- ? kstat_cpu(j).cpustat.nice : 0 );
|
|
|
+ j_dbs_info->prev_cpu_idle_up = get_cpu_idle_time(j);
|
|
|
j_dbs_info->prev_cpu_idle_down
|
|
|
= j_dbs_info->prev_cpu_idle_up;
|
|
|
}
|