sched/cpufreq: Change the worker kthread to SCHED_DEADLINE
Worker kthread needs to be able to change frequency for all other threads. Make it special, just under STOP class. Signed-off-by: Juri Lelli <juri.lelli@arm.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Claudio Scordino <claudio@evidence.eu.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Luca Abeni <luca.abeni@santannapisa.it> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Rafael J . Wysocki <rafael.j.wysocki@intel.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Viresh Kumar <viresh.kumar@linaro.org> Cc: alessio.balsini@arm.com Cc: bristot@redhat.com Cc: dietmar.eggemann@arm.com Cc: joelaf@google.com Cc: juri.lelli@redhat.com Cc: mathieu.poirier@linaro.org Cc: morten.rasmussen@arm.com Cc: patrick.bellasi@arm.com Cc: rjw@rjwysocki.net Cc: rostedt@goodmis.org Cc: tkjos@android.com Cc: tommaso.cucinotta@santannapisa.it Cc: vincent.guittot@linaro.org Link: http://lkml.kernel.org/r/20171204102325.5110-4-juri.lelli@redhat.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
@@ -78,7 +78,7 @@ static inline int dl_bw_cpus(int i)
|
||||
#endif
|
||||
|
||||
static inline
|
||||
void add_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
void __add_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
{
|
||||
u64 old = dl_rq->running_bw;
|
||||
|
||||
@@ -91,7 +91,7 @@ void add_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
}
|
||||
|
||||
static inline
|
||||
void sub_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
void __sub_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
{
|
||||
u64 old = dl_rq->running_bw;
|
||||
|
||||
@@ -105,7 +105,7 @@ void sub_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
}
|
||||
|
||||
static inline
|
||||
void add_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
void __add_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
{
|
||||
u64 old = dl_rq->this_bw;
|
||||
|
||||
@@ -115,7 +115,7 @@ void add_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
}
|
||||
|
||||
static inline
|
||||
void sub_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
void __sub_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
{
|
||||
u64 old = dl_rq->this_bw;
|
||||
|
||||
@@ -127,16 +127,46 @@ void sub_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
|
||||
SCHED_WARN_ON(dl_rq->running_bw > dl_rq->this_bw);
|
||||
}
|
||||
|
||||
static inline
|
||||
void add_rq_bw(struct sched_dl_entity *dl_se, struct dl_rq *dl_rq)
|
||||
{
|
||||
if (!dl_entity_is_special(dl_se))
|
||||
__add_rq_bw(dl_se->dl_bw, dl_rq);
|
||||
}
|
||||
|
||||
static inline
|
||||
void sub_rq_bw(struct sched_dl_entity *dl_se, struct dl_rq *dl_rq)
|
||||
{
|
||||
if (!dl_entity_is_special(dl_se))
|
||||
__sub_rq_bw(dl_se->dl_bw, dl_rq);
|
||||
}
|
||||
|
||||
static inline
|
||||
void add_running_bw(struct sched_dl_entity *dl_se, struct dl_rq *dl_rq)
|
||||
{
|
||||
if (!dl_entity_is_special(dl_se))
|
||||
__add_running_bw(dl_se->dl_bw, dl_rq);
|
||||
}
|
||||
|
||||
static inline
|
||||
void sub_running_bw(struct sched_dl_entity *dl_se, struct dl_rq *dl_rq)
|
||||
{
|
||||
if (!dl_entity_is_special(dl_se))
|
||||
__sub_running_bw(dl_se->dl_bw, dl_rq);
|
||||
}
|
||||
|
||||
void dl_change_utilization(struct task_struct *p, u64 new_bw)
|
||||
{
|
||||
struct rq *rq;
|
||||
|
||||
BUG_ON(p->dl.flags & SCHED_FLAG_SUGOV);
|
||||
|
||||
if (task_on_rq_queued(p))
|
||||
return;
|
||||
|
||||
rq = task_rq(p);
|
||||
if (p->dl.dl_non_contending) {
|
||||
sub_running_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_running_bw(&p->dl, &rq->dl);
|
||||
p->dl.dl_non_contending = 0;
|
||||
/*
|
||||
* If the timer handler is currently running and the
|
||||
@@ -148,8 +178,8 @@ void dl_change_utilization(struct task_struct *p, u64 new_bw)
|
||||
if (hrtimer_try_to_cancel(&p->dl.inactive_timer) == 1)
|
||||
put_task_struct(p);
|
||||
}
|
||||
sub_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
add_rq_bw(new_bw, &rq->dl);
|
||||
__sub_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
__add_rq_bw(new_bw, &rq->dl);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -221,6 +251,9 @@ static void task_non_contending(struct task_struct *p)
|
||||
if (dl_se->dl_runtime == 0)
|
||||
return;
|
||||
|
||||
if (dl_entity_is_special(dl_se))
|
||||
return;
|
||||
|
||||
WARN_ON(hrtimer_active(&dl_se->inactive_timer));
|
||||
WARN_ON(dl_se->dl_non_contending);
|
||||
|
||||
@@ -240,12 +273,12 @@ static void task_non_contending(struct task_struct *p)
|
||||
*/
|
||||
if (zerolag_time < 0) {
|
||||
if (dl_task(p))
|
||||
sub_running_bw(dl_se->dl_bw, dl_rq);
|
||||
sub_running_bw(dl_se, dl_rq);
|
||||
if (!dl_task(p) || p->state == TASK_DEAD) {
|
||||
struct dl_bw *dl_b = dl_bw_of(task_cpu(p));
|
||||
|
||||
if (p->state == TASK_DEAD)
|
||||
sub_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_rq_bw(&p->dl, &rq->dl);
|
||||
raw_spin_lock(&dl_b->lock);
|
||||
__dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p)));
|
||||
__dl_clear_params(p);
|
||||
@@ -272,7 +305,7 @@ static void task_contending(struct sched_dl_entity *dl_se, int flags)
|
||||
return;
|
||||
|
||||
if (flags & ENQUEUE_MIGRATED)
|
||||
add_rq_bw(dl_se->dl_bw, dl_rq);
|
||||
add_rq_bw(dl_se, dl_rq);
|
||||
|
||||
if (dl_se->dl_non_contending) {
|
||||
dl_se->dl_non_contending = 0;
|
||||
@@ -293,7 +326,7 @@ static void task_contending(struct sched_dl_entity *dl_se, int flags)
|
||||
* when the "inactive timer" fired).
|
||||
* So, add it back.
|
||||
*/
|
||||
add_running_bw(dl_se->dl_bw, dl_rq);
|
||||
add_running_bw(dl_se, dl_rq);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1149,6 +1182,9 @@ static void update_curr_dl(struct rq *rq)
|
||||
|
||||
sched_rt_avg_update(rq, delta_exec);
|
||||
|
||||
if (dl_entity_is_special(dl_se))
|
||||
return;
|
||||
|
||||
if (unlikely(dl_se->flags & SCHED_FLAG_RECLAIM))
|
||||
delta_exec = grub_reclaim(delta_exec, rq, &curr->dl);
|
||||
dl_se->runtime -= delta_exec;
|
||||
@@ -1211,8 +1247,8 @@ static enum hrtimer_restart inactive_task_timer(struct hrtimer *timer)
|
||||
struct dl_bw *dl_b = dl_bw_of(task_cpu(p));
|
||||
|
||||
if (p->state == TASK_DEAD && dl_se->dl_non_contending) {
|
||||
sub_running_bw(p->dl.dl_bw, dl_rq_of_se(&p->dl));
|
||||
sub_rq_bw(p->dl.dl_bw, dl_rq_of_se(&p->dl));
|
||||
sub_running_bw(&p->dl, dl_rq_of_se(&p->dl));
|
||||
sub_rq_bw(&p->dl, dl_rq_of_se(&p->dl));
|
||||
dl_se->dl_non_contending = 0;
|
||||
}
|
||||
|
||||
@@ -1229,7 +1265,7 @@ static enum hrtimer_restart inactive_task_timer(struct hrtimer *timer)
|
||||
sched_clock_tick();
|
||||
update_rq_clock(rq);
|
||||
|
||||
sub_running_bw(dl_se->dl_bw, &rq->dl);
|
||||
sub_running_bw(dl_se, &rq->dl);
|
||||
dl_se->dl_non_contending = 0;
|
||||
unlock:
|
||||
task_rq_unlock(rq, p, &rf);
|
||||
@@ -1423,8 +1459,8 @@ static void enqueue_task_dl(struct rq *rq, struct task_struct *p, int flags)
|
||||
dl_check_constrained_dl(&p->dl);
|
||||
|
||||
if (p->on_rq == TASK_ON_RQ_MIGRATING || flags & ENQUEUE_RESTORE) {
|
||||
add_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
add_running_bw(p->dl.dl_bw, &rq->dl);
|
||||
add_rq_bw(&p->dl, &rq->dl);
|
||||
add_running_bw(&p->dl, &rq->dl);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -1464,8 +1500,8 @@ static void dequeue_task_dl(struct rq *rq, struct task_struct *p, int flags)
|
||||
__dequeue_task_dl(rq, p, flags);
|
||||
|
||||
if (p->on_rq == TASK_ON_RQ_MIGRATING || flags & DEQUEUE_SAVE) {
|
||||
sub_running_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_running_bw(&p->dl, &rq->dl);
|
||||
sub_rq_bw(&p->dl, &rq->dl);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -1571,7 +1607,7 @@ static void migrate_task_rq_dl(struct task_struct *p)
|
||||
*/
|
||||
raw_spin_lock(&rq->lock);
|
||||
if (p->dl.dl_non_contending) {
|
||||
sub_running_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_running_bw(&p->dl, &rq->dl);
|
||||
p->dl.dl_non_contending = 0;
|
||||
/*
|
||||
* If the timer handler is currently running and the
|
||||
@@ -1583,7 +1619,7 @@ static void migrate_task_rq_dl(struct task_struct *p)
|
||||
if (hrtimer_try_to_cancel(&p->dl.inactive_timer) == 1)
|
||||
put_task_struct(p);
|
||||
}
|
||||
sub_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_rq_bw(&p->dl, &rq->dl);
|
||||
raw_spin_unlock(&rq->lock);
|
||||
}
|
||||
|
||||
@@ -2026,11 +2062,11 @@ retry:
|
||||
}
|
||||
|
||||
deactivate_task(rq, next_task, 0);
|
||||
sub_running_bw(next_task->dl.dl_bw, &rq->dl);
|
||||
sub_rq_bw(next_task->dl.dl_bw, &rq->dl);
|
||||
sub_running_bw(&next_task->dl, &rq->dl);
|
||||
sub_rq_bw(&next_task->dl, &rq->dl);
|
||||
set_task_cpu(next_task, later_rq->cpu);
|
||||
add_rq_bw(next_task->dl.dl_bw, &later_rq->dl);
|
||||
add_running_bw(next_task->dl.dl_bw, &later_rq->dl);
|
||||
add_rq_bw(&next_task->dl, &later_rq->dl);
|
||||
add_running_bw(&next_task->dl, &later_rq->dl);
|
||||
activate_task(later_rq, next_task, 0);
|
||||
ret = 1;
|
||||
|
||||
@@ -2118,11 +2154,11 @@ static void pull_dl_task(struct rq *this_rq)
|
||||
resched = true;
|
||||
|
||||
deactivate_task(src_rq, p, 0);
|
||||
sub_running_bw(p->dl.dl_bw, &src_rq->dl);
|
||||
sub_rq_bw(p->dl.dl_bw, &src_rq->dl);
|
||||
sub_running_bw(&p->dl, &src_rq->dl);
|
||||
sub_rq_bw(&p->dl, &src_rq->dl);
|
||||
set_task_cpu(p, this_cpu);
|
||||
add_rq_bw(p->dl.dl_bw, &this_rq->dl);
|
||||
add_running_bw(p->dl.dl_bw, &this_rq->dl);
|
||||
add_rq_bw(&p->dl, &this_rq->dl);
|
||||
add_running_bw(&p->dl, &this_rq->dl);
|
||||
activate_task(this_rq, p, 0);
|
||||
dmin = p->dl.deadline;
|
||||
|
||||
@@ -2231,7 +2267,7 @@ static void switched_from_dl(struct rq *rq, struct task_struct *p)
|
||||
task_non_contending(p);
|
||||
|
||||
if (!task_on_rq_queued(p))
|
||||
sub_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
sub_rq_bw(&p->dl, &rq->dl);
|
||||
|
||||
/*
|
||||
* We cannot use inactive_task_timer() to invoke sub_running_bw()
|
||||
@@ -2263,7 +2299,7 @@ static void switched_to_dl(struct rq *rq, struct task_struct *p)
|
||||
|
||||
/* If p is not queued we will update its parameters at next wakeup. */
|
||||
if (!task_on_rq_queued(p)) {
|
||||
add_rq_bw(p->dl.dl_bw, &rq->dl);
|
||||
add_rq_bw(&p->dl, &rq->dl);
|
||||
|
||||
return;
|
||||
}
|
||||
@@ -2442,6 +2478,9 @@ int sched_dl_overflow(struct task_struct *p, int policy,
|
||||
u64 new_bw = dl_policy(policy) ? to_ratio(period, runtime) : 0;
|
||||
int cpus, err = -1;
|
||||
|
||||
if (attr->sched_flags & SCHED_FLAG_SUGOV)
|
||||
return 0;
|
||||
|
||||
/* !deadline task may carry old deadline bandwidth */
|
||||
if (new_bw == p->dl.dl_bw && task_has_dl_policy(p))
|
||||
return 0;
|
||||
@@ -2528,6 +2567,10 @@ void __getparam_dl(struct task_struct *p, struct sched_attr *attr)
|
||||
*/
|
||||
bool __checkparam_dl(const struct sched_attr *attr)
|
||||
{
|
||||
/* special dl tasks don't actually use any parameter */
|
||||
if (attr->sched_flags & SCHED_FLAG_SUGOV)
|
||||
return true;
|
||||
|
||||
/* deadline != 0 */
|
||||
if (attr->sched_deadline == 0)
|
||||
return false;
|
||||
|
||||
Reference in New Issue
Block a user