unsigned long best_min_util = ULONG_MAX;
int best_cpu = -1;
- for_each_possible_cpu(cpu) {
+ for_each_cpu(cpu, cpu_active_mask) {
int i;
/* visit each coregroup only once */
util[cpu] += task_util_est(p);
}
- for_each_possible_cpu(cpu) {
+ for_each_cpu(cpu, cpu_active_mask) {
struct energy_table *table;
unsigned long max_util = 0, util_sum = 0;
unsigned long capacity;
if (!spin_trylock(&om_lock))
return;
- for_each_possible_cpu(cpu) {
+ for_each_cpu(cpu, cpu_active_mask) {
unsigned long flags;
struct rq *rq = cpu_rq(cpu);
struct sched_entity *se;
int i;
char state[30] = "prev_cpu";
- for_each_possible_cpu(cpu) {
+ for_each_cpu(cpu, cpu_active_mask) {
if (cpu != cpumask_first(cpu_coregroup_mask(cpu)))
continue;
unsigned long capacity_orig = capacity_orig_of(i);
unsigned long new_util, wake_util;
- if (!cpu_active(i))
- continue;
-
wake_util = cpu_util_wake(i, p);
new_util = wake_util + task_util_est(p);