if (cfs_rq_is_decayed(cfs_rq))
list_del_leaf_cfs_rq(cfs_rq);
}
- update_rt_rq_load_avg(rq_clock_task(rq), cpu, &rq->rt, 0);
#ifdef CONFIG_NO_HZ_COMMON
rq->last_blocked_load_update_tick = jiffies;
#endif
rq_lock_irqsave(rq, &rf);
update_rq_clock(rq);
update_cfs_rq_load_avg(cfs_rq_clock_task(cfs_rq), cfs_rq);
- update_rt_rq_load_avg(rq_clock_task(rq), cpu, &rq->rt, 0);
#ifdef CONFIG_NO_HZ_COMMON
rq->last_blocked_load_update_tick = jiffies;
#endif
u64 last_update_time;
last_update_time = rt_rq_last_update_time(rt_rq);
- update_rt_load_avg(last_update_time, rt_se);
+ __update_load_avg(last_update_time, cpu_of(rq_of_rt_rq(rt_rq)),
+ &rt_se->avg, 0, 0, NULL);
}
/*
update_curr_rt(rq);
- update_rt_rq_load_avg(rq_clock_task(rq), cpu_of(rq), &rq->rt, 1);
-
/*
* The previous task needs to be made eligible for pushing
* if it is still active
u64 now = rq_clock_task(rq);
update_curr_rt(rq);
- update_rt_rq_load_avg(now, cpu_of(rq), &rq->rt, 1);
for_each_sched_rt_entity(rt_se)
update_rt_load_avg(now, rt_se);