Revert "sched/pelt: Ensure that *_sum is always synced with *_avg"

This reverts commit 813ff24f1d.

Bug: 205915994
Test: build pass
Signed-off-by: Rick Yiu <rickyiu@google.com>
Change-Id: I77d9a103d71da43ce2fe9b630c7a5e12b62e8e42
This commit is contained in:
Rick Yiu
2021-11-16 12:33:07 +08:00
committed by Todd Kjos
parent 8630facf34
commit de0ba4ea3c

View File

@@ -3777,17 +3777,11 @@ static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *s
*/
static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
{
/*
* cfs_rq->avg.period_contrib can be used for both cfs_rq and se.
* See ___update_load_avg() for details.
*/
u32 divider = get_pelt_divider(&cfs_rq->avg);
dequeue_load_avg(cfs_rq, se);
sub_positive(&cfs_rq->avg.util_avg, se->avg.util_avg);
cfs_rq->avg.util_sum = cfs_rq->avg.util_avg * divider;
sub_positive(&cfs_rq->avg.util_sum, se->avg.util_sum);
sub_positive(&cfs_rq->avg.runnable_avg, se->avg.runnable_avg);
cfs_rq->avg.runnable_sum = cfs_rq->avg.runnable_avg * divider;
sub_positive(&cfs_rq->avg.runnable_sum, se->avg.runnable_sum);
add_tg_cfs_propagate(cfs_rq, -se->avg.load_sum);