cgroup/cpuset: Free DL BW in case can_attach() fails
commit 2ef269ef1ac006acf974793d975539244d77b28f upstream. cpuset_can_attach() can fail. Postpone DL BW allocation until all tasks have been checked. DL BW is not allocated per-task but as a sum over all DL tasks migrating. If multiple controllers are attached to the cgroup next to the cpuset controller a non-cpuset can_attach() can fail. In this case free DL BW in cpuset_cancel_attach(). Finally, update cpuset DL task count (nr_deadline_tasks) only in cpuset_attach(). Suggested-by: Waiman Long <longman@redhat.com> Signed-off-by: Dietmar Eggemann <dietmar.eggemann@arm.com> Signed-off-by: Juri Lelli <juri.lelli@redhat.com> Reviewed-by: Waiman Long <longman@redhat.com> Signed-off-by: Tejun Heo <tj@kernel.org> [ Fix conflicts in kernel/cgroup/cpuset.c due to new code being applied that is not applicable on this branch. Reject new code. ] Signed-off-by: Qais Yousef (Google) <qyousef@layalina.io> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:

committed by
Greg Kroah-Hartman

parent
4603c2a104
commit
2d69f68ad4
@@ -1657,7 +1657,7 @@ current_restore_flags(unsigned long orig_flags, unsigned long flags)
|
|||||||
}
|
}
|
||||||
|
|
||||||
extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
|
extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
|
||||||
extern int task_can_attach(struct task_struct *p, const struct cpumask *cs_effective_cpus);
|
extern int task_can_attach(struct task_struct *p);
|
||||||
extern int dl_bw_alloc(int cpu, u64 dl_bw);
|
extern int dl_bw_alloc(int cpu, u64 dl_bw);
|
||||||
extern void dl_bw_free(int cpu, u64 dl_bw);
|
extern void dl_bw_free(int cpu, u64 dl_bw);
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
|
@@ -167,6 +167,8 @@ struct cpuset {
|
|||||||
* know when to rebuild associated root domain bandwidth information.
|
* know when to rebuild associated root domain bandwidth information.
|
||||||
*/
|
*/
|
||||||
int nr_deadline_tasks;
|
int nr_deadline_tasks;
|
||||||
|
int nr_migrate_dl_tasks;
|
||||||
|
u64 sum_migrate_dl_bw;
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -2168,16 +2170,23 @@ static int fmeter_getrate(struct fmeter *fmp)
|
|||||||
|
|
||||||
static struct cpuset *cpuset_attach_old_cs;
|
static struct cpuset *cpuset_attach_old_cs;
|
||||||
|
|
||||||
|
static void reset_migrate_dl_data(struct cpuset *cs)
|
||||||
|
{
|
||||||
|
cs->nr_migrate_dl_tasks = 0;
|
||||||
|
cs->sum_migrate_dl_bw = 0;
|
||||||
|
}
|
||||||
|
|
||||||
/* Called by cgroups to determine if a cpuset is usable; cpuset_mutex held */
|
/* Called by cgroups to determine if a cpuset is usable; cpuset_mutex held */
|
||||||
static int cpuset_can_attach(struct cgroup_taskset *tset)
|
static int cpuset_can_attach(struct cgroup_taskset *tset)
|
||||||
{
|
{
|
||||||
struct cgroup_subsys_state *css;
|
struct cgroup_subsys_state *css;
|
||||||
struct cpuset *cs;
|
struct cpuset *cs, *oldcs;
|
||||||
struct task_struct *task;
|
struct task_struct *task;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
/* used later by cpuset_attach() */
|
/* used later by cpuset_attach() */
|
||||||
cpuset_attach_old_cs = task_cs(cgroup_taskset_first(tset, &css));
|
cpuset_attach_old_cs = task_cs(cgroup_taskset_first(tset, &css));
|
||||||
|
oldcs = cpuset_attach_old_cs;
|
||||||
cs = css_cs(css);
|
cs = css_cs(css);
|
||||||
|
|
||||||
mutex_lock(&cpuset_mutex);
|
mutex_lock(&cpuset_mutex);
|
||||||
@@ -2189,7 +2198,7 @@ static int cpuset_can_attach(struct cgroup_taskset *tset)
|
|||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
cgroup_taskset_for_each(task, css, tset) {
|
cgroup_taskset_for_each(task, css, tset) {
|
||||||
ret = task_can_attach(task, cs->effective_cpus);
|
ret = task_can_attach(task);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
ret = security_task_setscheduler(task);
|
ret = security_task_setscheduler(task);
|
||||||
@@ -2197,11 +2206,31 @@ static int cpuset_can_attach(struct cgroup_taskset *tset)
|
|||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
if (dl_task(task)) {
|
if (dl_task(task)) {
|
||||||
cs->nr_deadline_tasks++;
|
cs->nr_migrate_dl_tasks++;
|
||||||
cpuset_attach_old_cs->nr_deadline_tasks--;
|
cs->sum_migrate_dl_bw += task->dl.dl_bw;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!cs->nr_migrate_dl_tasks)
|
||||||
|
goto out_success;
|
||||||
|
|
||||||
|
if (!cpumask_intersects(oldcs->effective_cpus, cs->effective_cpus)) {
|
||||||
|
int cpu = cpumask_any_and(cpu_active_mask, cs->effective_cpus);
|
||||||
|
|
||||||
|
if (unlikely(cpu >= nr_cpu_ids)) {
|
||||||
|
reset_migrate_dl_data(cs);
|
||||||
|
ret = -EINVAL;
|
||||||
|
goto out_unlock;
|
||||||
|
}
|
||||||
|
|
||||||
|
ret = dl_bw_alloc(cpu, cs->sum_migrate_dl_bw);
|
||||||
|
if (ret) {
|
||||||
|
reset_migrate_dl_data(cs);
|
||||||
|
goto out_unlock;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
out_success:
|
||||||
/*
|
/*
|
||||||
* Mark attach is in progress. This makes validate_change() fail
|
* Mark attach is in progress. This makes validate_change() fail
|
||||||
* changes which zero cpus/mems_allowed.
|
* changes which zero cpus/mems_allowed.
|
||||||
@@ -2225,6 +2254,14 @@ static void cpuset_cancel_attach(struct cgroup_taskset *tset)
|
|||||||
cs->attach_in_progress--;
|
cs->attach_in_progress--;
|
||||||
if (!cs->attach_in_progress)
|
if (!cs->attach_in_progress)
|
||||||
wake_up(&cpuset_attach_wq);
|
wake_up(&cpuset_attach_wq);
|
||||||
|
|
||||||
|
if (cs->nr_migrate_dl_tasks) {
|
||||||
|
int cpu = cpumask_any(cs->effective_cpus);
|
||||||
|
|
||||||
|
dl_bw_free(cpu, cs->sum_migrate_dl_bw);
|
||||||
|
reset_migrate_dl_data(cs);
|
||||||
|
}
|
||||||
|
|
||||||
mutex_unlock(&cpuset_mutex);
|
mutex_unlock(&cpuset_mutex);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2299,6 +2336,12 @@ static void cpuset_attach(struct cgroup_taskset *tset)
|
|||||||
|
|
||||||
cs->old_mems_allowed = cpuset_attach_nodemask_to;
|
cs->old_mems_allowed = cpuset_attach_nodemask_to;
|
||||||
|
|
||||||
|
if (cs->nr_migrate_dl_tasks) {
|
||||||
|
cs->nr_deadline_tasks += cs->nr_migrate_dl_tasks;
|
||||||
|
oldcs->nr_deadline_tasks -= cs->nr_migrate_dl_tasks;
|
||||||
|
reset_migrate_dl_data(cs);
|
||||||
|
}
|
||||||
|
|
||||||
cs->attach_in_progress--;
|
cs->attach_in_progress--;
|
||||||
if (!cs->attach_in_progress)
|
if (!cs->attach_in_progress)
|
||||||
wake_up(&cpuset_attach_wq);
|
wake_up(&cpuset_attach_wq);
|
||||||
|
@@ -6600,8 +6600,7 @@ int cpuset_cpumask_can_shrink(const struct cpumask *cur,
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
int task_can_attach(struct task_struct *p,
|
int task_can_attach(struct task_struct *p)
|
||||||
const struct cpumask *cs_effective_cpus)
|
|
||||||
{
|
{
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
@@ -6614,21 +6613,9 @@ int task_can_attach(struct task_struct *p,
|
|||||||
* success of set_cpus_allowed_ptr() on all attached tasks
|
* success of set_cpus_allowed_ptr() on all attached tasks
|
||||||
* before cpus_mask may be changed.
|
* before cpus_mask may be changed.
|
||||||
*/
|
*/
|
||||||
if (p->flags & PF_NO_SETAFFINITY) {
|
if (p->flags & PF_NO_SETAFFINITY)
|
||||||
ret = -EINVAL;
|
ret = -EINVAL;
|
||||||
goto out;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (dl_task(p) && !cpumask_intersects(task_rq(p)->rd->span,
|
|
||||||
cs_effective_cpus)) {
|
|
||||||
int cpu = cpumask_any_and(cpu_active_mask, cs_effective_cpus);
|
|
||||||
|
|
||||||
if (unlikely(cpu >= nr_cpu_ids))
|
|
||||||
return -EINVAL;
|
|
||||||
ret = dl_bw_alloc(cpu, p->dl.dl_bw);
|
|
||||||
}
|
|
||||||
|
|
||||||
out:
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user