Merge branch 'cpus4096-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'cpus4096-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: (31 commits) NR_CPUS: Replace NR_CPUS in speedstep-centrino.c cpumask: Provide a generic set of CPUMASK_ALLOC macros, FIXUP NR_CPUS: Replace NR_CPUS in cpufreq userspace routines NR_CPUS: Replace per_cpu(..., smp_processor_id()) with __get_cpu_var NR_CPUS: Replace NR_CPUS in arch/x86/kernel/genapic_flat_64.c NR_CPUS: Replace NR_CPUS in arch/x86/kernel/genx2apic_uv_x.c NR_CPUS: Replace NR_CPUS in arch/x86/kernel/cpu/proc.c NR_CPUS: Replace NR_CPUS in arch/x86/kernel/cpu/mcheck/mce_64.c cpumask: Optimize cpumask_of_cpu in lib/smp_processor_id.c, fix cpumask: Use optimized CPUMASK_ALLOC macros in the centrino_target cpumask: Provide a generic set of CPUMASK_ALLOC macros cpumask: Optimize cpumask_of_cpu in lib/smp_processor_id.c cpumask: Optimize cpumask_of_cpu in kernel/time/tick-common.c cpumask: Optimize cpumask_of_cpu in drivers/misc/sgi-xp/xpc_main.c cpumask: Optimize cpumask_of_cpu in arch/x86/kernel/ldt.c cpumask: Optimize cpumask_of_cpu in arch/x86/kernel/io_apic_64.c cpumask: Replace cpumask_of_cpu with cpumask_of_cpu_ptr Revert "cpumask: introduce new APIs" cpumask: make for_each_cpu_mask a bit smaller net: Pass reference to cpumask variable in net/sunrpc/svc.c ... Fix up trivial conflicts in drivers/cpufreq/cpufreq.c manually
This commit is contained in:
@@ -413,7 +413,7 @@ void __ref enable_nonboot_cpus(void)
|
||||
goto out;
|
||||
|
||||
printk("Enabling non-boot CPUs ...\n");
|
||||
for_each_cpu_mask(cpu, frozen_cpus) {
|
||||
for_each_cpu_mask_nr(cpu, frozen_cpus) {
|
||||
error = _cpu_up(cpu, 1);
|
||||
if (!error) {
|
||||
printk("CPU%d is up\n", cpu);
|
||||
|
@@ -106,7 +106,7 @@ static void force_quiescent_state(struct rcu_data *rdp,
|
||||
*/
|
||||
cpus_and(cpumask, rcp->cpumask, cpu_online_map);
|
||||
cpu_clear(rdp->cpu, cpumask);
|
||||
for_each_cpu_mask(cpu, cpumask)
|
||||
for_each_cpu_mask_nr(cpu, cpumask)
|
||||
smp_send_reschedule(cpu);
|
||||
}
|
||||
}
|
||||
|
@@ -756,7 +756,7 @@ rcu_try_flip_idle(void)
|
||||
|
||||
/* Now ask each CPU for acknowledgement of the flip. */
|
||||
|
||||
for_each_cpu_mask(cpu, rcu_cpu_online_map) {
|
||||
for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) {
|
||||
per_cpu(rcu_flip_flag, cpu) = rcu_flipped;
|
||||
dyntick_save_progress_counter(cpu);
|
||||
}
|
||||
@@ -774,7 +774,7 @@ rcu_try_flip_waitack(void)
|
||||
int cpu;
|
||||
|
||||
RCU_TRACE_ME(rcupreempt_trace_try_flip_a1);
|
||||
for_each_cpu_mask(cpu, rcu_cpu_online_map)
|
||||
for_each_cpu_mask_nr(cpu, rcu_cpu_online_map)
|
||||
if (rcu_try_flip_waitack_needed(cpu) &&
|
||||
per_cpu(rcu_flip_flag, cpu) != rcu_flip_seen) {
|
||||
RCU_TRACE_ME(rcupreempt_trace_try_flip_ae1);
|
||||
@@ -806,7 +806,7 @@ rcu_try_flip_waitzero(void)
|
||||
/* Check to see if the sum of the "last" counters is zero. */
|
||||
|
||||
RCU_TRACE_ME(rcupreempt_trace_try_flip_z1);
|
||||
for_each_cpu_mask(cpu, rcu_cpu_online_map)
|
||||
for_each_cpu_mask_nr(cpu, rcu_cpu_online_map)
|
||||
sum += RCU_DATA_CPU(cpu)->rcu_flipctr[lastidx];
|
||||
if (sum != 0) {
|
||||
RCU_TRACE_ME(rcupreempt_trace_try_flip_ze1);
|
||||
@@ -821,7 +821,7 @@ rcu_try_flip_waitzero(void)
|
||||
smp_mb(); /* ^^^^^^^^^^^^ */
|
||||
|
||||
/* Call for a memory barrier from each CPU. */
|
||||
for_each_cpu_mask(cpu, rcu_cpu_online_map) {
|
||||
for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) {
|
||||
per_cpu(rcu_mb_flag, cpu) = rcu_mb_needed;
|
||||
dyntick_save_progress_counter(cpu);
|
||||
}
|
||||
@@ -841,7 +841,7 @@ rcu_try_flip_waitmb(void)
|
||||
int cpu;
|
||||
|
||||
RCU_TRACE_ME(rcupreempt_trace_try_flip_m1);
|
||||
for_each_cpu_mask(cpu, rcu_cpu_online_map)
|
||||
for_each_cpu_mask_nr(cpu, rcu_cpu_online_map)
|
||||
if (rcu_try_flip_waitmb_needed(cpu) &&
|
||||
per_cpu(rcu_mb_flag, cpu) != rcu_mb_done) {
|
||||
RCU_TRACE_ME(rcupreempt_trace_try_flip_me1);
|
||||
|
@@ -2108,7 +2108,7 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
|
||||
/* Tally up the load of all CPUs in the group */
|
||||
avg_load = 0;
|
||||
|
||||
for_each_cpu_mask(i, group->cpumask) {
|
||||
for_each_cpu_mask_nr(i, group->cpumask) {
|
||||
/* Bias balancing toward cpus of our domain */
|
||||
if (local_group)
|
||||
load = source_load(i, load_idx);
|
||||
@@ -2150,7 +2150,7 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu,
|
||||
/* Traverse only the allowed CPUs */
|
||||
cpus_and(*tmp, group->cpumask, p->cpus_allowed);
|
||||
|
||||
for_each_cpu_mask(i, *tmp) {
|
||||
for_each_cpu_mask_nr(i, *tmp) {
|
||||
load = weighted_cpuload(i);
|
||||
|
||||
if (load < min_load || (load == min_load && i == this_cpu)) {
|
||||
@@ -3168,7 +3168,7 @@ find_busiest_group(struct sched_domain *sd, int this_cpu,
|
||||
max_cpu_load = 0;
|
||||
min_cpu_load = ~0UL;
|
||||
|
||||
for_each_cpu_mask(i, group->cpumask) {
|
||||
for_each_cpu_mask_nr(i, group->cpumask) {
|
||||
struct rq *rq;
|
||||
|
||||
if (!cpu_isset(i, *cpus))
|
||||
@@ -3447,7 +3447,7 @@ find_busiest_queue(struct sched_group *group, enum cpu_idle_type idle,
|
||||
unsigned long max_load = 0;
|
||||
int i;
|
||||
|
||||
for_each_cpu_mask(i, group->cpumask) {
|
||||
for_each_cpu_mask_nr(i, group->cpumask) {
|
||||
unsigned long wl;
|
||||
|
||||
if (!cpu_isset(i, *cpus))
|
||||
@@ -3989,7 +3989,7 @@ static void run_rebalance_domains(struct softirq_action *h)
|
||||
int balance_cpu;
|
||||
|
||||
cpu_clear(this_cpu, cpus);
|
||||
for_each_cpu_mask(balance_cpu, cpus) {
|
||||
for_each_cpu_mask_nr(balance_cpu, cpus) {
|
||||
/*
|
||||
* If this cpu gets work to do, stop the load balancing
|
||||
* work being done for other cpus. Next load
|
||||
@@ -6802,7 +6802,7 @@ init_sched_build_groups(const cpumask_t *span, const cpumask_t *cpu_map,
|
||||
|
||||
cpus_clear(*covered);
|
||||
|
||||
for_each_cpu_mask(i, *span) {
|
||||
for_each_cpu_mask_nr(i, *span) {
|
||||
struct sched_group *sg;
|
||||
int group = group_fn(i, cpu_map, &sg, tmpmask);
|
||||
int j;
|
||||
@@ -6813,7 +6813,7 @@ init_sched_build_groups(const cpumask_t *span, const cpumask_t *cpu_map,
|
||||
cpus_clear(sg->cpumask);
|
||||
sg->__cpu_power = 0;
|
||||
|
||||
for_each_cpu_mask(j, *span) {
|
||||
for_each_cpu_mask_nr(j, *span) {
|
||||
if (group_fn(j, cpu_map, NULL, tmpmask) != group)
|
||||
continue;
|
||||
|
||||
@@ -7013,7 +7013,7 @@ static void init_numa_sched_groups_power(struct sched_group *group_head)
|
||||
if (!sg)
|
||||
return;
|
||||
do {
|
||||
for_each_cpu_mask(j, sg->cpumask) {
|
||||
for_each_cpu_mask_nr(j, sg->cpumask) {
|
||||
struct sched_domain *sd;
|
||||
|
||||
sd = &per_cpu(phys_domains, j);
|
||||
@@ -7038,7 +7038,7 @@ static void free_sched_groups(const cpumask_t *cpu_map, cpumask_t *nodemask)
|
||||
{
|
||||
int cpu, i;
|
||||
|
||||
for_each_cpu_mask(cpu, *cpu_map) {
|
||||
for_each_cpu_mask_nr(cpu, *cpu_map) {
|
||||
struct sched_group **sched_group_nodes
|
||||
= sched_group_nodes_bycpu[cpu];
|
||||
|
||||
@@ -7277,7 +7277,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map,
|
||||
/*
|
||||
* Set up domains for cpus specified by the cpu_map.
|
||||
*/
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
struct sched_domain *sd = NULL, *p;
|
||||
SCHED_CPUMASK_VAR(nodemask, allmasks);
|
||||
|
||||
@@ -7344,7 +7344,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map,
|
||||
|
||||
#ifdef CONFIG_SCHED_SMT
|
||||
/* Set up CPU (sibling) groups */
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
SCHED_CPUMASK_VAR(this_sibling_map, allmasks);
|
||||
SCHED_CPUMASK_VAR(send_covered, allmasks);
|
||||
|
||||
@@ -7361,7 +7361,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map,
|
||||
|
||||
#ifdef CONFIG_SCHED_MC
|
||||
/* Set up multi-core groups */
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
SCHED_CPUMASK_VAR(this_core_map, allmasks);
|
||||
SCHED_CPUMASK_VAR(send_covered, allmasks);
|
||||
|
||||
@@ -7428,7 +7428,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map,
|
||||
goto error;
|
||||
}
|
||||
sched_group_nodes[i] = sg;
|
||||
for_each_cpu_mask(j, *nodemask) {
|
||||
for_each_cpu_mask_nr(j, *nodemask) {
|
||||
struct sched_domain *sd;
|
||||
|
||||
sd = &per_cpu(node_domains, j);
|
||||
@@ -7474,21 +7474,21 @@ static int __build_sched_domains(const cpumask_t *cpu_map,
|
||||
|
||||
/* Calculate CPU power for physical packages and nodes */
|
||||
#ifdef CONFIG_SCHED_SMT
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
struct sched_domain *sd = &per_cpu(cpu_domains, i);
|
||||
|
||||
init_sched_groups_power(i, sd);
|
||||
}
|
||||
#endif
|
||||
#ifdef CONFIG_SCHED_MC
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
struct sched_domain *sd = &per_cpu(core_domains, i);
|
||||
|
||||
init_sched_groups_power(i, sd);
|
||||
}
|
||||
#endif
|
||||
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
struct sched_domain *sd = &per_cpu(phys_domains, i);
|
||||
|
||||
init_sched_groups_power(i, sd);
|
||||
@@ -7508,7 +7508,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map,
|
||||
#endif
|
||||
|
||||
/* Attach the domains */
|
||||
for_each_cpu_mask(i, *cpu_map) {
|
||||
for_each_cpu_mask_nr(i, *cpu_map) {
|
||||
struct sched_domain *sd;
|
||||
#ifdef CONFIG_SCHED_SMT
|
||||
sd = &per_cpu(cpu_domains, i);
|
||||
@@ -7603,7 +7603,7 @@ static void detach_destroy_domains(const cpumask_t *cpu_map)
|
||||
|
||||
unregister_sched_domain_sysctl();
|
||||
|
||||
for_each_cpu_mask(i, *cpu_map)
|
||||
for_each_cpu_mask_nr(i, *cpu_map)
|
||||
cpu_attach_domain(NULL, &def_root_domain, i);
|
||||
synchronize_sched();
|
||||
arch_destroy_sched_domains(cpu_map, &tmpmask);
|
||||
|
@@ -1031,7 +1031,7 @@ static int wake_idle(int cpu, struct task_struct *p)
|
||||
|| ((sd->flags & SD_WAKE_IDLE_FAR)
|
||||
&& !task_hot(p, task_rq(p)->clock, sd))) {
|
||||
cpus_and(tmp, sd->span, p->cpus_allowed);
|
||||
for_each_cpu_mask(i, tmp) {
|
||||
for_each_cpu_mask_nr(i, tmp) {
|
||||
if (idle_cpu(i)) {
|
||||
if (i != task_cpu(p)) {
|
||||
schedstat_inc(p,
|
||||
|
@@ -240,7 +240,7 @@ static int do_balance_runtime(struct rt_rq *rt_rq)
|
||||
|
||||
spin_lock(&rt_b->rt_runtime_lock);
|
||||
rt_period = ktime_to_ns(rt_b->rt_period);
|
||||
for_each_cpu_mask(i, rd->span) {
|
||||
for_each_cpu_mask_nr(i, rd->span) {
|
||||
struct rt_rq *iter = sched_rt_period_rt_rq(rt_b, i);
|
||||
s64 diff;
|
||||
|
||||
@@ -1107,7 +1107,7 @@ static int pull_rt_task(struct rq *this_rq)
|
||||
|
||||
next = pick_next_task_rt(this_rq);
|
||||
|
||||
for_each_cpu_mask(cpu, this_rq->rd->rto_mask) {
|
||||
for_each_cpu_mask_nr(cpu, this_rq->rd->rto_mask) {
|
||||
if (this_cpu == cpu)
|
||||
continue;
|
||||
|
||||
|
@@ -33,8 +33,9 @@ static int stopmachine(void *cpu)
|
||||
{
|
||||
int irqs_disabled = 0;
|
||||
int prepared = 0;
|
||||
cpumask_of_cpu_ptr(cpumask, (int)(long)cpu);
|
||||
|
||||
set_cpus_allowed_ptr(current, &cpumask_of_cpu((int)(long)cpu));
|
||||
set_cpus_allowed_ptr(current, cpumask);
|
||||
|
||||
/* Ack: we are alive */
|
||||
smp_mb(); /* Theoretically the ack = 0 might not be on this CPU yet. */
|
||||
|
@@ -301,7 +301,7 @@ static int add_del_listener(pid_t pid, cpumask_t *maskp, int isadd)
|
||||
return -EINVAL;
|
||||
|
||||
if (isadd == REGISTER) {
|
||||
for_each_cpu_mask(cpu, mask) {
|
||||
for_each_cpu_mask_nr(cpu, mask) {
|
||||
s = kmalloc_node(sizeof(struct listener), GFP_KERNEL,
|
||||
cpu_to_node(cpu));
|
||||
if (!s)
|
||||
@@ -320,7 +320,7 @@ static int add_del_listener(pid_t pid, cpumask_t *maskp, int isadd)
|
||||
|
||||
/* Deregister or cleanup */
|
||||
cleanup:
|
||||
for_each_cpu_mask(cpu, mask) {
|
||||
for_each_cpu_mask_nr(cpu, mask) {
|
||||
listeners = &per_cpu(listener_array, cpu);
|
||||
down_write(&listeners->sem);
|
||||
list_for_each_entry_safe(s, tmp, &listeners->list, list) {
|
||||
|
@@ -145,9 +145,9 @@ static void clocksource_watchdog(unsigned long data)
|
||||
* Cycle through CPUs to check if the CPUs stay
|
||||
* synchronized to each other.
|
||||
*/
|
||||
int next_cpu = next_cpu(raw_smp_processor_id(), cpu_online_map);
|
||||
int next_cpu = next_cpu_nr(raw_smp_processor_id(), cpu_online_map);
|
||||
|
||||
if (next_cpu >= NR_CPUS)
|
||||
if (next_cpu >= nr_cpu_ids)
|
||||
next_cpu = first_cpu(cpu_online_map);
|
||||
watchdog_timer.expires += WATCHDOG_INTERVAL;
|
||||
add_timer_on(&watchdog_timer, next_cpu);
|
||||
|
@@ -399,8 +399,7 @@ again:
|
||||
mask = CPU_MASK_NONE;
|
||||
now = ktime_get();
|
||||
/* Find all expired events */
|
||||
for (cpu = first_cpu(tick_broadcast_oneshot_mask); cpu != NR_CPUS;
|
||||
cpu = next_cpu(cpu, tick_broadcast_oneshot_mask)) {
|
||||
for_each_cpu_mask_nr(cpu, tick_broadcast_oneshot_mask) {
|
||||
td = &per_cpu(tick_cpu_device, cpu);
|
||||
if (td->evtdev->next_event.tv64 <= now.tv64)
|
||||
cpu_set(cpu, mask);
|
||||
|
@@ -135,7 +135,7 @@ void tick_setup_periodic(struct clock_event_device *dev, int broadcast)
|
||||
*/
|
||||
static void tick_setup_device(struct tick_device *td,
|
||||
struct clock_event_device *newdev, int cpu,
|
||||
cpumask_t cpumask)
|
||||
const cpumask_t *cpumask)
|
||||
{
|
||||
ktime_t next_event;
|
||||
void (*handler)(struct clock_event_device *) = NULL;
|
||||
@@ -169,8 +169,8 @@ static void tick_setup_device(struct tick_device *td,
|
||||
* When the device is not per cpu, pin the interrupt to the
|
||||
* current cpu:
|
||||
*/
|
||||
if (!cpus_equal(newdev->cpumask, cpumask))
|
||||
irq_set_affinity(newdev->irq, cpumask);
|
||||
if (!cpus_equal(newdev->cpumask, *cpumask))
|
||||
irq_set_affinity(newdev->irq, *cpumask);
|
||||
|
||||
/*
|
||||
* When global broadcasting is active, check if the current
|
||||
@@ -196,20 +196,20 @@ static int tick_check_new_device(struct clock_event_device *newdev)
|
||||
struct tick_device *td;
|
||||
int cpu, ret = NOTIFY_OK;
|
||||
unsigned long flags;
|
||||
cpumask_t cpumask;
|
||||
cpumask_of_cpu_ptr_declare(cpumask);
|
||||
|
||||
spin_lock_irqsave(&tick_device_lock, flags);
|
||||
|
||||
cpu = smp_processor_id();
|
||||
cpumask_of_cpu_ptr_next(cpumask, cpu);
|
||||
if (!cpu_isset(cpu, newdev->cpumask))
|
||||
goto out_bc;
|
||||
|
||||
td = &per_cpu(tick_cpu_device, cpu);
|
||||
curdev = td->evtdev;
|
||||
cpumask = cpumask_of_cpu(cpu);
|
||||
|
||||
/* cpu local device ? */
|
||||
if (!cpus_equal(newdev->cpumask, cpumask)) {
|
||||
if (!cpus_equal(newdev->cpumask, *cpumask)) {
|
||||
|
||||
/*
|
||||
* If the cpu affinity of the device interrupt can not
|
||||
@@ -222,7 +222,7 @@ static int tick_check_new_device(struct clock_event_device *newdev)
|
||||
* If we have a cpu local device already, do not replace it
|
||||
* by a non cpu local device
|
||||
*/
|
||||
if (curdev && cpus_equal(curdev->cpumask, cpumask))
|
||||
if (curdev && cpus_equal(curdev->cpumask, *cpumask))
|
||||
goto out_bc;
|
||||
}
|
||||
|
||||
|
@@ -213,7 +213,9 @@ static void start_stack_timers(void)
|
||||
int cpu;
|
||||
|
||||
for_each_online_cpu(cpu) {
|
||||
set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu));
|
||||
cpumask_of_cpu_ptr(new_mask, cpu);
|
||||
|
||||
set_cpus_allowed_ptr(current, new_mask);
|
||||
start_stack_timer(cpu);
|
||||
}
|
||||
set_cpus_allowed_ptr(current, &saved_mask);
|
||||
|
@@ -397,7 +397,7 @@ void flush_workqueue(struct workqueue_struct *wq)
|
||||
might_sleep();
|
||||
lock_acquire(&wq->lockdep_map, 0, 0, 0, 2, _THIS_IP_);
|
||||
lock_release(&wq->lockdep_map, 1, _THIS_IP_);
|
||||
for_each_cpu_mask(cpu, *cpu_map)
|
||||
for_each_cpu_mask_nr(cpu, *cpu_map)
|
||||
flush_cpu_workqueue(per_cpu_ptr(wq->cpu_wq, cpu));
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(flush_workqueue);
|
||||
@@ -477,7 +477,7 @@ static void wait_on_work(struct work_struct *work)
|
||||
wq = cwq->wq;
|
||||
cpu_map = wq_cpu_map(wq);
|
||||
|
||||
for_each_cpu_mask(cpu, *cpu_map)
|
||||
for_each_cpu_mask_nr(cpu, *cpu_map)
|
||||
wait_on_cpu_work(per_cpu_ptr(wq->cpu_wq, cpu), work);
|
||||
}
|
||||
|
||||
@@ -813,7 +813,7 @@ void destroy_workqueue(struct workqueue_struct *wq)
|
||||
list_del(&wq->list);
|
||||
spin_unlock(&workqueue_lock);
|
||||
|
||||
for_each_cpu_mask(cpu, *cpu_map)
|
||||
for_each_cpu_mask_nr(cpu, *cpu_map)
|
||||
cleanup_workqueue_thread(per_cpu_ptr(wq->cpu_wq, cpu));
|
||||
put_online_cpus();
|
||||
|
||||
|
Reference in New Issue
Block a user