Merge branch 'cpumask-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
Conflicts: arch/x86/include/asm/topology.h drivers/oprofile/buffer_sync.c (Both cases: changed in Linus' tree, removed in Ingo's).
This commit is contained in:
@@ -26,12 +26,12 @@ static int bigsmp_apic_id_registered(void)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static const cpumask_t *bigsmp_target_cpus(void)
|
||||
static const struct cpumask *bigsmp_target_cpus(void)
|
||||
{
|
||||
#ifdef CONFIG_SMP
|
||||
return &cpu_online_map;
|
||||
return cpu_online_mask;
|
||||
#else
|
||||
return &cpumask_of_cpu(0);
|
||||
return cpumask_of(0);
|
||||
#endif
|
||||
}
|
||||
|
||||
@@ -118,9 +118,9 @@ static int bigsmp_check_phys_apicid_present(int boot_cpu_physical_apicid)
|
||||
}
|
||||
|
||||
/* As we are using single CPU as destination, pick only one CPU here */
|
||||
static unsigned int bigsmp_cpu_mask_to_apicid(const cpumask_t *cpumask)
|
||||
static unsigned int bigsmp_cpu_mask_to_apicid(const struct cpumask *cpumask)
|
||||
{
|
||||
return bigsmp_cpu_to_logical_apicid(first_cpu(*cpumask));
|
||||
return bigsmp_cpu_to_logical_apicid(cpumask_first(cpumask));
|
||||
}
|
||||
|
||||
static unsigned int bigsmp_cpu_mask_to_apicid_and(const struct cpumask *cpumask,
|
||||
@@ -188,10 +188,10 @@ static const struct dmi_system_id bigsmp_dmi_table[] = {
|
||||
{ } /* NULL entry stops DMI scanning */
|
||||
};
|
||||
|
||||
static void bigsmp_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
static void bigsmp_vector_allocation_domain(int cpu, struct cpumask *retmask)
|
||||
{
|
||||
cpus_clear(*retmask);
|
||||
cpu_set(cpu, *retmask);
|
||||
cpumask_clear(retmask);
|
||||
cpumask_set_cpu(cpu, retmask);
|
||||
}
|
||||
|
||||
static int probe_bigsmp(void)
|
||||
|
@@ -410,7 +410,7 @@ static void es7000_enable_apic_mode(void)
|
||||
WARN(1, "Command failed, status = %x\n", mip_status);
|
||||
}
|
||||
|
||||
static void es7000_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
static void es7000_vector_allocation_domain(int cpu, struct cpumask *retmask)
|
||||
{
|
||||
/* Careful. Some cpus do not strictly honor the set of cpus
|
||||
* specified in the interrupt destination when using lowest
|
||||
@@ -420,7 +420,8 @@ static void es7000_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
* deliver interrupts to the wrong hyperthread when only one
|
||||
* hyperthread was specified in the interrupt desitination.
|
||||
*/
|
||||
*retmask = (cpumask_t){ { [0] = APIC_ALL_CPUS, } };
|
||||
cpumask_clear(retmask);
|
||||
cpumask_bits(retmask)[0] = APIC_ALL_CPUS;
|
||||
}
|
||||
|
||||
|
||||
@@ -455,14 +456,14 @@ static int es7000_apic_id_registered(void)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static const cpumask_t *target_cpus_cluster(void)
|
||||
static const struct cpumask *target_cpus_cluster(void)
|
||||
{
|
||||
return &CPU_MASK_ALL;
|
||||
return cpu_all_mask;
|
||||
}
|
||||
|
||||
static const cpumask_t *es7000_target_cpus(void)
|
||||
static const struct cpumask *es7000_target_cpus(void)
|
||||
{
|
||||
return &cpumask_of_cpu(smp_processor_id());
|
||||
return cpumask_of(smp_processor_id());
|
||||
}
|
||||
|
||||
static unsigned long
|
||||
@@ -517,7 +518,7 @@ static void es7000_setup_apic_routing(void)
|
||||
"Enabling APIC mode: %s. Using %d I/O APICs, target cpus %lx\n",
|
||||
(apic_version[apic] == 0x14) ?
|
||||
"Physical Cluster" : "Logical Cluster",
|
||||
nr_ioapics, cpus_addr(*es7000_target_cpus())[0]);
|
||||
nr_ioapics, cpumask_bits(es7000_target_cpus())[0]);
|
||||
}
|
||||
|
||||
static int es7000_apicid_to_node(int logical_apicid)
|
||||
@@ -572,7 +573,7 @@ static int es7000_check_phys_apicid_present(int cpu_physical_apicid)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static unsigned int es7000_cpu_mask_to_apicid(const cpumask_t *cpumask)
|
||||
static unsigned int es7000_cpu_mask_to_apicid(const struct cpumask *cpumask)
|
||||
{
|
||||
unsigned int round = 0;
|
||||
int cpu, uninitialized_var(apicid);
|
||||
|
@@ -39,7 +39,7 @@
|
||||
int unknown_nmi_panic;
|
||||
int nmi_watchdog_enabled;
|
||||
|
||||
static cpumask_t backtrace_mask = CPU_MASK_NONE;
|
||||
static cpumask_var_t backtrace_mask;
|
||||
|
||||
/* nmi_active:
|
||||
* >0: the lapic NMI watchdog is active, but can be disabled
|
||||
@@ -138,6 +138,7 @@ int __init check_nmi_watchdog(void)
|
||||
if (!prev_nmi_count)
|
||||
goto error;
|
||||
|
||||
alloc_cpumask_var(&backtrace_mask, GFP_KERNEL);
|
||||
printk(KERN_INFO "Testing NMI watchdog ... ");
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
@@ -413,14 +414,14 @@ nmi_watchdog_tick(struct pt_regs *regs, unsigned reason)
|
||||
touched = 1;
|
||||
}
|
||||
|
||||
if (cpu_isset(cpu, backtrace_mask)) {
|
||||
if (cpumask_test_cpu(cpu, backtrace_mask)) {
|
||||
static DEFINE_SPINLOCK(lock); /* Serialise the printks */
|
||||
|
||||
spin_lock(&lock);
|
||||
printk(KERN_WARNING "NMI backtrace for cpu %d\n", cpu);
|
||||
dump_stack();
|
||||
spin_unlock(&lock);
|
||||
cpu_clear(cpu, backtrace_mask);
|
||||
cpumask_clear_cpu(cpu, backtrace_mask);
|
||||
}
|
||||
|
||||
/* Could check oops_in_progress here too, but it's safer not to */
|
||||
@@ -554,10 +555,10 @@ void __trigger_all_cpu_backtrace(void)
|
||||
{
|
||||
int i;
|
||||
|
||||
backtrace_mask = cpu_online_map;
|
||||
cpumask_copy(backtrace_mask, cpu_online_mask);
|
||||
/* Wait for up to 10 seconds for all CPUs to do the backtrace */
|
||||
for (i = 0; i < 10 * 1000; i++) {
|
||||
if (cpus_empty(backtrace_mask))
|
||||
if (cpumask_empty(backtrace_mask))
|
||||
break;
|
||||
mdelay(1);
|
||||
}
|
||||
|
@@ -334,9 +334,9 @@ static inline void numaq_smp_callin_clear_local_apic(void)
|
||||
clear_local_APIC();
|
||||
}
|
||||
|
||||
static inline const cpumask_t *numaq_target_cpus(void)
|
||||
static inline const struct cpumask *numaq_target_cpus(void)
|
||||
{
|
||||
return &CPU_MASK_ALL;
|
||||
return cpu_all_mask;
|
||||
}
|
||||
|
||||
static inline unsigned long
|
||||
@@ -427,7 +427,7 @@ static inline int numaq_check_phys_apicid_present(int boot_cpu_physical_apicid)
|
||||
* We use physical apicids here, not logical, so just return the default
|
||||
* physical broadcast to stop people from breaking us
|
||||
*/
|
||||
static inline unsigned int numaq_cpu_mask_to_apicid(const cpumask_t *cpumask)
|
||||
static unsigned int numaq_cpu_mask_to_apicid(const struct cpumask *cpumask)
|
||||
{
|
||||
return 0x0F;
|
||||
}
|
||||
@@ -462,7 +462,7 @@ static int probe_numaq(void)
|
||||
return found_numaq;
|
||||
}
|
||||
|
||||
static void numaq_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
static void numaq_vector_allocation_domain(int cpu, struct cpumask *retmask)
|
||||
{
|
||||
/* Careful. Some cpus do not strictly honor the set of cpus
|
||||
* specified in the interrupt destination when using lowest
|
||||
@@ -472,7 +472,8 @@ static void numaq_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
* deliver interrupts to the wrong hyperthread when only one
|
||||
* hyperthread was specified in the interrupt desitination.
|
||||
*/
|
||||
*retmask = (cpumask_t){ { [0] = APIC_ALL_CPUS, } };
|
||||
cpumask_clear(retmask);
|
||||
cpumask_bits(retmask)[0] = APIC_ALL_CPUS;
|
||||
}
|
||||
|
||||
static void numaq_setup_portio_remap(void)
|
||||
|
@@ -83,7 +83,8 @@ static void default_vector_allocation_domain(int cpu, struct cpumask *retmask)
|
||||
* deliver interrupts to the wrong hyperthread when only one
|
||||
* hyperthread was specified in the interrupt desitination.
|
||||
*/
|
||||
*retmask = (cpumask_t) { { [0] = APIC_ALL_CPUS } };
|
||||
cpumask_clear(retmask);
|
||||
cpumask_bits(retmask)[0] = APIC_ALL_CPUS;
|
||||
}
|
||||
|
||||
/* should be called last. */
|
||||
|
@@ -53,23 +53,19 @@ static unsigned summit_get_apic_id(unsigned long x)
|
||||
return (x >> 24) & 0xFF;
|
||||
}
|
||||
|
||||
static inline void summit_send_IPI_mask(const cpumask_t *mask, int vector)
|
||||
static inline void summit_send_IPI_mask(const struct cpumask *mask, int vector)
|
||||
{
|
||||
default_send_IPI_mask_sequence_logical(mask, vector);
|
||||
}
|
||||
|
||||
static void summit_send_IPI_allbutself(int vector)
|
||||
{
|
||||
cpumask_t mask = cpu_online_map;
|
||||
cpu_clear(smp_processor_id(), mask);
|
||||
|
||||
if (!cpus_empty(mask))
|
||||
summit_send_IPI_mask(&mask, vector);
|
||||
default_send_IPI_mask_allbutself_logical(cpu_online_mask, vector);
|
||||
}
|
||||
|
||||
static void summit_send_IPI_all(int vector)
|
||||
{
|
||||
summit_send_IPI_mask(&cpu_online_map, vector);
|
||||
summit_send_IPI_mask(cpu_online_mask, vector);
|
||||
}
|
||||
|
||||
#include <asm/tsc.h>
|
||||
@@ -186,13 +182,13 @@ static inline int is_WPEG(struct rio_detail *rio){
|
||||
|
||||
#define SUMMIT_APIC_DFR_VALUE (APIC_DFR_CLUSTER)
|
||||
|
||||
static const cpumask_t *summit_target_cpus(void)
|
||||
static const struct cpumask *summit_target_cpus(void)
|
||||
{
|
||||
/* CPU_MASK_ALL (0xff) has undefined behaviour with
|
||||
* dest_LowestPrio mode logical clustered apic interrupt routing
|
||||
* Just start on cpu 0. IRQ balancing will spread load
|
||||
*/
|
||||
return &cpumask_of_cpu(0);
|
||||
return cpumask_of(0);
|
||||
}
|
||||
|
||||
static unsigned long summit_check_apicid_used(physid_mask_t bitmap, int apicid)
|
||||
@@ -289,7 +285,7 @@ static int summit_check_phys_apicid_present(int boot_cpu_physical_apicid)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static unsigned int summit_cpu_mask_to_apicid(const cpumask_t *cpumask)
|
||||
static unsigned int summit_cpu_mask_to_apicid(const struct cpumask *cpumask)
|
||||
{
|
||||
unsigned int round = 0;
|
||||
int cpu, apicid = 0;
|
||||
@@ -346,7 +342,7 @@ static int probe_summit(void)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void summit_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
static void summit_vector_allocation_domain(int cpu, struct cpumask *retmask)
|
||||
{
|
||||
/* Careful. Some cpus do not strictly honor the set of cpus
|
||||
* specified in the interrupt destination when using lowest
|
||||
@@ -356,7 +352,8 @@ static void summit_vector_allocation_domain(int cpu, cpumask_t *retmask)
|
||||
* deliver interrupts to the wrong hyperthread when only one
|
||||
* hyperthread was specified in the interrupt desitination.
|
||||
*/
|
||||
*retmask = (cpumask_t){ { [0] = APIC_ALL_CPUS, } };
|
||||
cpumask_clear(retmask);
|
||||
cpumask_bits(retmask)[0] = APIC_ALL_CPUS;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_X86_SUMMIT_NUMA
|
||||
|
@@ -466,7 +466,7 @@ static const lookup_t error_table[] = {
|
||||
* @err: APM BIOS return code
|
||||
*
|
||||
* Write a meaningful log entry to the kernel log in the event of
|
||||
* an APM error.
|
||||
* an APM error. Note that this also handles (negative) kernel errors.
|
||||
*/
|
||||
|
||||
static void apm_error(char *str, int err)
|
||||
@@ -478,42 +478,13 @@ static void apm_error(char *str, int err)
|
||||
break;
|
||||
if (i < ERROR_COUNT)
|
||||
printk(KERN_NOTICE "apm: %s: %s\n", str, error_table[i].msg);
|
||||
else if (err < 0)
|
||||
printk(KERN_NOTICE "apm: %s: linux error code %i\n", str, err);
|
||||
else
|
||||
printk(KERN_NOTICE "apm: %s: unknown error code %#2.2x\n",
|
||||
str, err);
|
||||
}
|
||||
|
||||
/*
|
||||
* Lock APM functionality to physical CPU 0
|
||||
*/
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
|
||||
static cpumask_t apm_save_cpus(void)
|
||||
{
|
||||
cpumask_t x = current->cpus_allowed;
|
||||
/* Some bioses don't like being called from CPU != 0 */
|
||||
set_cpus_allowed(current, cpumask_of_cpu(0));
|
||||
BUG_ON(smp_processor_id() != 0);
|
||||
return x;
|
||||
}
|
||||
|
||||
static inline void apm_restore_cpus(cpumask_t mask)
|
||||
{
|
||||
set_cpus_allowed(current, mask);
|
||||
}
|
||||
|
||||
#else
|
||||
|
||||
/*
|
||||
* No CPU lockdown needed on a uniprocessor
|
||||
*/
|
||||
|
||||
#define apm_save_cpus() (current->cpus_allowed)
|
||||
#define apm_restore_cpus(x) (void)(x)
|
||||
|
||||
#endif
|
||||
|
||||
/*
|
||||
* These are the actual BIOS calls. Depending on APM_ZERO_SEGS and
|
||||
* apm_info.allow_ints, we are being really paranoid here! Not only
|
||||
@@ -568,16 +539,23 @@ static inline void apm_irq_restore(unsigned long flags)
|
||||
# define APM_DO_RESTORE_SEGS
|
||||
#endif
|
||||
|
||||
struct apm_bios_call {
|
||||
u32 func;
|
||||
/* In and out */
|
||||
u32 ebx;
|
||||
u32 ecx;
|
||||
/* Out only */
|
||||
u32 eax;
|
||||
u32 edx;
|
||||
u32 esi;
|
||||
|
||||
/* Error: -ENOMEM, or bits 8-15 of eax */
|
||||
int err;
|
||||
};
|
||||
|
||||
/**
|
||||
* apm_bios_call - Make an APM BIOS 32bit call
|
||||
* @func: APM function to execute
|
||||
* @ebx_in: EBX register for call entry
|
||||
* @ecx_in: ECX register for call entry
|
||||
* @eax: EAX register return
|
||||
* @ebx: EBX register return
|
||||
* @ecx: ECX register return
|
||||
* @edx: EDX register return
|
||||
* @esi: ESI register return
|
||||
* __apm_bios_call - Make an APM BIOS 32bit call
|
||||
* @_call: pointer to struct apm_bios_call.
|
||||
*
|
||||
* Make an APM call using the 32bit protected mode interface. The
|
||||
* caller is responsible for knowing if APM BIOS is configured and
|
||||
@@ -586,35 +564,109 @@ static inline void apm_irq_restore(unsigned long flags)
|
||||
* flag is loaded into AL. If there is an error, then the error
|
||||
* code is returned in AH (bits 8-15 of eax) and this function
|
||||
* returns non-zero.
|
||||
*
|
||||
* Note: this makes the call on the current CPU.
|
||||
*/
|
||||
|
||||
static u8 apm_bios_call(u32 func, u32 ebx_in, u32 ecx_in,
|
||||
u32 *eax, u32 *ebx, u32 *ecx, u32 *edx, u32 *esi)
|
||||
static long __apm_bios_call(void *_call)
|
||||
{
|
||||
APM_DECL_SEGS
|
||||
unsigned long flags;
|
||||
cpumask_t cpus;
|
||||
int cpu;
|
||||
struct desc_struct save_desc_40;
|
||||
struct desc_struct *gdt;
|
||||
|
||||
cpus = apm_save_cpus();
|
||||
struct apm_bios_call *call = _call;
|
||||
|
||||
cpu = get_cpu();
|
||||
BUG_ON(cpu != 0);
|
||||
gdt = get_cpu_gdt_table(cpu);
|
||||
save_desc_40 = gdt[0x40 / 8];
|
||||
gdt[0x40 / 8] = bad_bios_desc;
|
||||
|
||||
apm_irq_save(flags);
|
||||
APM_DO_SAVE_SEGS;
|
||||
apm_bios_call_asm(func, ebx_in, ecx_in, eax, ebx, ecx, edx, esi);
|
||||
apm_bios_call_asm(call->func, call->ebx, call->ecx,
|
||||
&call->eax, &call->ebx, &call->ecx, &call->edx,
|
||||
&call->esi);
|
||||
APM_DO_RESTORE_SEGS;
|
||||
apm_irq_restore(flags);
|
||||
gdt[0x40 / 8] = save_desc_40;
|
||||
put_cpu();
|
||||
apm_restore_cpus(cpus);
|
||||
|
||||
return *eax & 0xff;
|
||||
return call->eax & 0xff;
|
||||
}
|
||||
|
||||
/* Run __apm_bios_call or __apm_bios_call_simple on CPU 0 */
|
||||
static int on_cpu0(long (*fn)(void *), struct apm_bios_call *call)
|
||||
{
|
||||
int ret;
|
||||
|
||||
/* Don't bother with work_on_cpu in the common case, so we don't
|
||||
* have to worry about OOM or overhead. */
|
||||
if (get_cpu() == 0) {
|
||||
ret = fn(call);
|
||||
put_cpu();
|
||||
} else {
|
||||
put_cpu();
|
||||
ret = work_on_cpu(0, fn, call);
|
||||
}
|
||||
|
||||
/* work_on_cpu can fail with -ENOMEM */
|
||||
if (ret < 0)
|
||||
call->err = ret;
|
||||
else
|
||||
call->err = (call->eax >> 8) & 0xff;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
* apm_bios_call - Make an APM BIOS 32bit call (on CPU 0)
|
||||
* @call: the apm_bios_call registers.
|
||||
*
|
||||
* If there is an error, it is returned in @call.err.
|
||||
*/
|
||||
static int apm_bios_call(struct apm_bios_call *call)
|
||||
{
|
||||
return on_cpu0(__apm_bios_call, call);
|
||||
}
|
||||
|
||||
/**
|
||||
* __apm_bios_call_simple - Make an APM BIOS 32bit call (on CPU 0)
|
||||
* @_call: pointer to struct apm_bios_call.
|
||||
*
|
||||
* Make a BIOS call that returns one value only, or just status.
|
||||
* If there is an error, then the error code is returned in AH
|
||||
* (bits 8-15 of eax) and this function returns non-zero (it can
|
||||
* also return -ENOMEM). This is used for simpler BIOS operations.
|
||||
* This call may hold interrupts off for a long time on some laptops.
|
||||
*
|
||||
* Note: this makes the call on the current CPU.
|
||||
*/
|
||||
static long __apm_bios_call_simple(void *_call)
|
||||
{
|
||||
u8 error;
|
||||
APM_DECL_SEGS
|
||||
unsigned long flags;
|
||||
int cpu;
|
||||
struct desc_struct save_desc_40;
|
||||
struct desc_struct *gdt;
|
||||
struct apm_bios_call *call = _call;
|
||||
|
||||
cpu = get_cpu();
|
||||
BUG_ON(cpu != 0);
|
||||
gdt = get_cpu_gdt_table(cpu);
|
||||
save_desc_40 = gdt[0x40 / 8];
|
||||
gdt[0x40 / 8] = bad_bios_desc;
|
||||
|
||||
apm_irq_save(flags);
|
||||
APM_DO_SAVE_SEGS;
|
||||
error = apm_bios_call_simple_asm(call->func, call->ebx, call->ecx,
|
||||
&call->eax);
|
||||
APM_DO_RESTORE_SEGS;
|
||||
apm_irq_restore(flags);
|
||||
gdt[0x40 / 8] = save_desc_40;
|
||||
put_cpu();
|
||||
return error;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -623,40 +675,28 @@ static u8 apm_bios_call(u32 func, u32 ebx_in, u32 ecx_in,
|
||||
* @ebx_in: EBX register value for BIOS call
|
||||
* @ecx_in: ECX register value for BIOS call
|
||||
* @eax: EAX register on return from the BIOS call
|
||||
* @err: bits
|
||||
*
|
||||
* Make a BIOS call that returns one value only, or just status.
|
||||
* If there is an error, then the error code is returned in AH
|
||||
* (bits 8-15 of eax) and this function returns non-zero. This is
|
||||
* used for simpler BIOS operations. This call may hold interrupts
|
||||
* off for a long time on some laptops.
|
||||
* If there is an error, then the error code is returned in @err
|
||||
* and this function returns non-zero. This is used for simpler
|
||||
* BIOS operations. This call may hold interrupts off for a long
|
||||
* time on some laptops.
|
||||
*/
|
||||
|
||||
static u8 apm_bios_call_simple(u32 func, u32 ebx_in, u32 ecx_in, u32 *eax)
|
||||
static int apm_bios_call_simple(u32 func, u32 ebx_in, u32 ecx_in, u32 *eax,
|
||||
int *err)
|
||||
{
|
||||
u8 error;
|
||||
APM_DECL_SEGS
|
||||
unsigned long flags;
|
||||
cpumask_t cpus;
|
||||
int cpu;
|
||||
struct desc_struct save_desc_40;
|
||||
struct desc_struct *gdt;
|
||||
struct apm_bios_call call;
|
||||
int ret;
|
||||
|
||||
cpus = apm_save_cpus();
|
||||
call.func = func;
|
||||
call.ebx = ebx_in;
|
||||
call.ecx = ecx_in;
|
||||
|
||||
cpu = get_cpu();
|
||||
gdt = get_cpu_gdt_table(cpu);
|
||||
save_desc_40 = gdt[0x40 / 8];
|
||||
gdt[0x40 / 8] = bad_bios_desc;
|
||||
|
||||
apm_irq_save(flags);
|
||||
APM_DO_SAVE_SEGS;
|
||||
error = apm_bios_call_simple_asm(func, ebx_in, ecx_in, eax);
|
||||
APM_DO_RESTORE_SEGS;
|
||||
apm_irq_restore(flags);
|
||||
gdt[0x40 / 8] = save_desc_40;
|
||||
put_cpu();
|
||||
apm_restore_cpus(cpus);
|
||||
return error;
|
||||
ret = on_cpu0(__apm_bios_call_simple, &call);
|
||||
*eax = call.eax;
|
||||
*err = call.err;
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -678,9 +718,10 @@ static u8 apm_bios_call_simple(u32 func, u32 ebx_in, u32 ecx_in, u32 *eax)
|
||||
static int apm_driver_version(u_short *val)
|
||||
{
|
||||
u32 eax;
|
||||
int err;
|
||||
|
||||
if (apm_bios_call_simple(APM_FUNC_VERSION, 0, *val, &eax))
|
||||
return (eax >> 8) & 0xff;
|
||||
if (apm_bios_call_simple(APM_FUNC_VERSION, 0, *val, &eax, &err))
|
||||
return err;
|
||||
*val = eax;
|
||||
return APM_SUCCESS;
|
||||
}
|
||||
@@ -701,22 +742,21 @@ static int apm_driver_version(u_short *val)
|
||||
* that APM 1.2 is in use. If no messges are pending the value 0x80
|
||||
* is returned (No power management events pending).
|
||||
*/
|
||||
|
||||
static int apm_get_event(apm_event_t *event, apm_eventinfo_t *info)
|
||||
{
|
||||
u32 eax;
|
||||
u32 ebx;
|
||||
u32 ecx;
|
||||
u32 dummy;
|
||||
struct apm_bios_call call;
|
||||
|
||||
if (apm_bios_call(APM_FUNC_GET_EVENT, 0, 0, &eax, &ebx, &ecx,
|
||||
&dummy, &dummy))
|
||||
return (eax >> 8) & 0xff;
|
||||
*event = ebx;
|
||||
call.func = APM_FUNC_GET_EVENT;
|
||||
call.ebx = call.ecx = 0;
|
||||
|
||||
if (apm_bios_call(&call))
|
||||
return call.err;
|
||||
|
||||
*event = call.ebx;
|
||||
if (apm_info.connection_version < 0x0102)
|
||||
*info = ~0; /* indicate info not valid */
|
||||
else
|
||||
*info = ecx;
|
||||
*info = call.ecx;
|
||||
return APM_SUCCESS;
|
||||
}
|
||||
|
||||
@@ -737,9 +777,10 @@ static int apm_get_event(apm_event_t *event, apm_eventinfo_t *info)
|
||||
static int set_power_state(u_short what, u_short state)
|
||||
{
|
||||
u32 eax;
|
||||
int err;
|
||||
|
||||
if (apm_bios_call_simple(APM_FUNC_SET_STATE, what, state, &eax))
|
||||
return (eax >> 8) & 0xff;
|
||||
if (apm_bios_call_simple(APM_FUNC_SET_STATE, what, state, &eax, &err))
|
||||
return err;
|
||||
return APM_SUCCESS;
|
||||
}
|
||||
|
||||
@@ -770,6 +811,7 @@ static int apm_do_idle(void)
|
||||
u8 ret = 0;
|
||||
int idled = 0;
|
||||
int polling;
|
||||
int err;
|
||||
|
||||
polling = !!(current_thread_info()->status & TS_POLLING);
|
||||
if (polling) {
|
||||
@@ -782,7 +824,7 @@ static int apm_do_idle(void)
|
||||
}
|
||||
if (!need_resched()) {
|
||||
idled = 1;
|
||||
ret = apm_bios_call_simple(APM_FUNC_IDLE, 0, 0, &eax);
|
||||
ret = apm_bios_call_simple(APM_FUNC_IDLE, 0, 0, &eax, &err);
|
||||
}
|
||||
if (polling)
|
||||
current_thread_info()->status |= TS_POLLING;
|
||||
@@ -797,8 +839,7 @@ static int apm_do_idle(void)
|
||||
* Only report the failure the first 5 times.
|
||||
*/
|
||||
if (++t < 5) {
|
||||
printk(KERN_DEBUG "apm_do_idle failed (%d)\n",
|
||||
(eax >> 8) & 0xff);
|
||||
printk(KERN_DEBUG "apm_do_idle failed (%d)\n", err);
|
||||
t = jiffies;
|
||||
}
|
||||
return -1;
|
||||
@@ -816,9 +857,10 @@ static int apm_do_idle(void)
|
||||
static void apm_do_busy(void)
|
||||
{
|
||||
u32 dummy;
|
||||
int err;
|
||||
|
||||
if (clock_slowed || ALWAYS_CALL_BUSY) {
|
||||
(void)apm_bios_call_simple(APM_FUNC_BUSY, 0, 0, &dummy);
|
||||
(void)apm_bios_call_simple(APM_FUNC_BUSY, 0, 0, &dummy, &err);
|
||||
clock_slowed = 0;
|
||||
}
|
||||
}
|
||||
@@ -937,7 +979,7 @@ static void apm_power_off(void)
|
||||
|
||||
/* Some bioses don't like being called from CPU != 0 */
|
||||
if (apm_info.realmode_power_off) {
|
||||
(void)apm_save_cpus();
|
||||
set_cpus_allowed_ptr(current, cpumask_of(0));
|
||||
machine_real_restart(po_bios_call, sizeof(po_bios_call));
|
||||
} else {
|
||||
(void)set_system_power_state(APM_STATE_OFF);
|
||||
@@ -956,12 +998,13 @@ static void apm_power_off(void)
|
||||
static int apm_enable_power_management(int enable)
|
||||
{
|
||||
u32 eax;
|
||||
int err;
|
||||
|
||||
if ((enable == 0) && (apm_info.bios.flags & APM_BIOS_DISENGAGED))
|
||||
return APM_NOT_ENGAGED;
|
||||
if (apm_bios_call_simple(APM_FUNC_ENABLE_PM, APM_DEVICE_BALL,
|
||||
enable, &eax))
|
||||
return (eax >> 8) & 0xff;
|
||||
enable, &eax, &err))
|
||||
return err;
|
||||
if (enable)
|
||||
apm_info.bios.flags &= ~APM_BIOS_DISABLED;
|
||||
else
|
||||
@@ -986,24 +1029,23 @@ static int apm_enable_power_management(int enable)
|
||||
|
||||
static int apm_get_power_status(u_short *status, u_short *bat, u_short *life)
|
||||
{
|
||||
u32 eax;
|
||||
u32 ebx;
|
||||
u32 ecx;
|
||||
u32 edx;
|
||||
u32 dummy;
|
||||
struct apm_bios_call call;
|
||||
|
||||
call.func = APM_FUNC_GET_STATUS;
|
||||
call.ebx = APM_DEVICE_ALL;
|
||||
call.ecx = 0;
|
||||
|
||||
if (apm_info.get_power_status_broken)
|
||||
return APM_32_UNSUPPORTED;
|
||||
if (apm_bios_call(APM_FUNC_GET_STATUS, APM_DEVICE_ALL, 0,
|
||||
&eax, &ebx, &ecx, &edx, &dummy))
|
||||
return (eax >> 8) & 0xff;
|
||||
*status = ebx;
|
||||
*bat = ecx;
|
||||
if (apm_bios_call(&call))
|
||||
return call.err;
|
||||
*status = call.ebx;
|
||||
*bat = call.ecx;
|
||||
if (apm_info.get_power_status_swabinminutes) {
|
||||
*life = swab16((u16)edx);
|
||||
*life = swab16((u16)call.edx);
|
||||
*life |= 0x8000;
|
||||
} else
|
||||
*life = edx;
|
||||
*life = call.edx;
|
||||
return APM_SUCCESS;
|
||||
}
|
||||
|
||||
@@ -1048,12 +1090,14 @@ static int apm_get_battery_status(u_short which, u_short *status,
|
||||
static int apm_engage_power_management(u_short device, int enable)
|
||||
{
|
||||
u32 eax;
|
||||
int err;
|
||||
|
||||
if ((enable == 0) && (device == APM_DEVICE_ALL)
|
||||
&& (apm_info.bios.flags & APM_BIOS_DISABLED))
|
||||
return APM_DISABLED;
|
||||
if (apm_bios_call_simple(APM_FUNC_ENGAGE_PM, device, enable, &eax))
|
||||
return (eax >> 8) & 0xff;
|
||||
if (apm_bios_call_simple(APM_FUNC_ENGAGE_PM, device, enable,
|
||||
&eax, &err))
|
||||
return err;
|
||||
if (device == APM_DEVICE_ALL) {
|
||||
if (enable)
|
||||
apm_info.bios.flags &= ~APM_BIOS_DISENGAGED;
|
||||
@@ -1689,16 +1733,14 @@ static int apm(void *unused)
|
||||
char *power_stat;
|
||||
char *bat_stat;
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
/* 2002/08/01 - WT
|
||||
* This is to avoid random crashes at boot time during initialization
|
||||
* on SMP systems in case of "apm=power-off" mode. Seen on ASUS A7M266D.
|
||||
* Some bioses don't like being called from CPU != 0.
|
||||
* Method suggested by Ingo Molnar.
|
||||
*/
|
||||
set_cpus_allowed(current, cpumask_of_cpu(0));
|
||||
set_cpus_allowed_ptr(current, cpumask_of(0));
|
||||
BUG_ON(smp_processor_id() != 0);
|
||||
#endif
|
||||
|
||||
if (apm_info.connection_version == 0) {
|
||||
apm_info.connection_version = apm_info.bios.version;
|
||||
|
@@ -41,8 +41,6 @@
|
||||
|
||||
#include "cpu.h"
|
||||
|
||||
#ifdef CONFIG_X86_64
|
||||
|
||||
/* all of these masks are initialized in setup_cpu_local_masks() */
|
||||
cpumask_var_t cpu_initialized_mask;
|
||||
cpumask_var_t cpu_callout_mask;
|
||||
@@ -60,16 +58,6 @@ void __init setup_cpu_local_masks(void)
|
||||
alloc_bootmem_cpumask_var(&cpu_sibling_setup_mask);
|
||||
}
|
||||
|
||||
#else /* CONFIG_X86_32 */
|
||||
|
||||
cpumask_t cpu_sibling_setup_map;
|
||||
cpumask_t cpu_callout_map;
|
||||
cpumask_t cpu_initialized;
|
||||
cpumask_t cpu_callin_map;
|
||||
|
||||
#endif /* CONFIG_X86_32 */
|
||||
|
||||
|
||||
static const struct cpu_dev *this_cpu __cpuinitdata;
|
||||
|
||||
DEFINE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page) = { .gdt = {
|
||||
@@ -859,6 +847,7 @@ static void vgetcpu_set_mode(void)
|
||||
void __init identify_boot_cpu(void)
|
||||
{
|
||||
identify_cpu(&boot_cpu_data);
|
||||
init_c1e_mask();
|
||||
#ifdef CONFIG_X86_32
|
||||
sysenter_setup();
|
||||
enable_sep_cpu();
|
||||
|
@@ -211,7 +211,7 @@ static int cpufreq_p4_cpu_init(struct cpufreq_policy *policy)
|
||||
unsigned int i;
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
cpumask_copy(policy->cpus, &per_cpu(cpu_sibling_map, policy->cpu));
|
||||
cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu));
|
||||
#endif
|
||||
|
||||
/* Errata workaround */
|
||||
|
@@ -54,7 +54,10 @@ static DEFINE_PER_CPU(struct powernow_k8_data *, powernow_data);
|
||||
static int cpu_family = CPU_OPTERON;
|
||||
|
||||
#ifndef CONFIG_SMP
|
||||
DEFINE_PER_CPU(cpumask_t, cpu_core_map);
|
||||
static inline const struct cpumask *cpu_core_mask(int cpu)
|
||||
{
|
||||
return cpumask_of(0);
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Return a frequency in MHz, given an input fid */
|
||||
@@ -699,7 +702,7 @@ static int fill_powernow_table(struct powernow_k8_data *data,
|
||||
|
||||
dprintk("cfid 0x%x, cvid 0x%x\n", data->currfid, data->currvid);
|
||||
data->powernow_table = powernow_table;
|
||||
if (first_cpu(per_cpu(cpu_core_map, data->cpu)) == data->cpu)
|
||||
if (cpumask_first(cpu_core_mask(data->cpu)) == data->cpu)
|
||||
print_basics(data);
|
||||
|
||||
for (j = 0; j < data->numps; j++)
|
||||
@@ -862,7 +865,7 @@ static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data)
|
||||
|
||||
/* fill in data */
|
||||
data->numps = data->acpi_data.state_count;
|
||||
if (first_cpu(per_cpu(cpu_core_map, data->cpu)) == data->cpu)
|
||||
if (cpumask_first(cpu_core_mask(data->cpu)) == data->cpu)
|
||||
print_basics(data);
|
||||
powernow_k8_acpi_pst_values(data, 0);
|
||||
|
||||
@@ -1300,7 +1303,7 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol)
|
||||
if (cpu_family == CPU_HW_PSTATE)
|
||||
cpumask_copy(pol->cpus, cpumask_of(pol->cpu));
|
||||
else
|
||||
cpumask_copy(pol->cpus, &per_cpu(cpu_core_map, pol->cpu));
|
||||
cpumask_copy(pol->cpus, cpu_core_mask(pol->cpu));
|
||||
data->available_cores = pol->cpus;
|
||||
|
||||
if (cpu_family == CPU_HW_PSTATE)
|
||||
@@ -1365,7 +1368,7 @@ static unsigned int powernowk8_get(unsigned int cpu)
|
||||
unsigned int khz = 0;
|
||||
unsigned int first;
|
||||
|
||||
first = first_cpu(per_cpu(cpu_core_map, cpu));
|
||||
first = cpumask_first(cpu_core_mask(cpu));
|
||||
data = per_cpu(powernow_data, first);
|
||||
|
||||
if (!data)
|
||||
|
@@ -321,7 +321,7 @@ static int speedstep_cpu_init(struct cpufreq_policy *policy)
|
||||
|
||||
/* only run on CPU to be set, or on its sibling */
|
||||
#ifdef CONFIG_SMP
|
||||
cpumask_copy(policy->cpus, &per_cpu(cpu_sibling_map, policy->cpu));
|
||||
cpumask_copy(policy->cpus, cpu_sibling_mask(policy->cpu));
|
||||
#endif
|
||||
|
||||
cpus_allowed = current->cpus_allowed;
|
||||
|
@@ -159,7 +159,7 @@ struct _cpuid4_info_regs {
|
||||
unsigned long can_disable;
|
||||
};
|
||||
|
||||
#ifdef CONFIG_PCI
|
||||
#if defined(CONFIG_PCI) && defined(CONFIG_SYSFS)
|
||||
static struct pci_device_id k8_nb_id[] = {
|
||||
{ PCI_DEVICE(PCI_VENDOR_ID_AMD, 0x1103) },
|
||||
{ PCI_DEVICE(PCI_VENDOR_ID_AMD, 0x1203) },
|
||||
@@ -324,15 +324,6 @@ __cpuinit cpuid4_cache_lookup_regs(int index,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
__cpuinit cpuid4_cache_lookup(int index, struct _cpuid4_info *this_leaf)
|
||||
{
|
||||
struct _cpuid4_info_regs *leaf_regs =
|
||||
(struct _cpuid4_info_regs *)this_leaf;
|
||||
|
||||
return cpuid4_cache_lookup_regs(index, leaf_regs);
|
||||
}
|
||||
|
||||
static int __cpuinit find_num_cache_leaves(void)
|
||||
{
|
||||
unsigned int eax, ebx, ecx, edx;
|
||||
@@ -508,6 +499,8 @@ unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
|
||||
return l2;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_SYSFS
|
||||
|
||||
/* pointer to _cpuid4_info array (for each cache leaf) */
|
||||
static DEFINE_PER_CPU(struct _cpuid4_info *, cpuid4_info);
|
||||
#define CPUID4_INFO_IDX(x, y) (&((per_cpu(cpuid4_info, x))[y]))
|
||||
@@ -571,6 +564,15 @@ static void __cpuinit free_cache_attributes(unsigned int cpu)
|
||||
per_cpu(cpuid4_info, cpu) = NULL;
|
||||
}
|
||||
|
||||
static int
|
||||
__cpuinit cpuid4_cache_lookup(int index, struct _cpuid4_info *this_leaf)
|
||||
{
|
||||
struct _cpuid4_info_regs *leaf_regs =
|
||||
(struct _cpuid4_info_regs *)this_leaf;
|
||||
|
||||
return cpuid4_cache_lookup_regs(index, leaf_regs);
|
||||
}
|
||||
|
||||
static void __cpuinit get_cpu_leaves(void *_retval)
|
||||
{
|
||||
int j, *retval = _retval, cpu = smp_processor_id();
|
||||
@@ -612,8 +614,6 @@ static int __cpuinit detect_cache_attributes(unsigned int cpu)
|
||||
return retval;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_SYSFS
|
||||
|
||||
#include <linux/kobject.h>
|
||||
#include <linux/sysfs.h>
|
||||
|
||||
|
@@ -990,7 +990,7 @@ static struct sysdev_attribute *mce_attributes[] = {
|
||||
NULL
|
||||
};
|
||||
|
||||
static cpumask_t mce_device_initialized = CPU_MASK_NONE;
|
||||
static cpumask_var_t mce_device_initialized;
|
||||
|
||||
/* Per cpu sysdev init. All of the cpus still share the same ctl bank */
|
||||
static __cpuinit int mce_create_device(unsigned int cpu)
|
||||
@@ -1021,7 +1021,7 @@ static __cpuinit int mce_create_device(unsigned int cpu)
|
||||
if (err)
|
||||
goto error2;
|
||||
}
|
||||
cpu_set(cpu, mce_device_initialized);
|
||||
cpumask_set_cpu(cpu, mce_device_initialized);
|
||||
|
||||
return 0;
|
||||
error2:
|
||||
@@ -1043,7 +1043,7 @@ static __cpuinit void mce_remove_device(unsigned int cpu)
|
||||
{
|
||||
int i;
|
||||
|
||||
if (!cpu_isset(cpu, mce_device_initialized))
|
||||
if (!cpumask_test_cpu(cpu, mce_device_initialized))
|
||||
return;
|
||||
|
||||
for (i = 0; mce_attributes[i]; i++)
|
||||
@@ -1053,7 +1053,7 @@ static __cpuinit void mce_remove_device(unsigned int cpu)
|
||||
sysdev_remove_file(&per_cpu(device_mce, cpu),
|
||||
&bank_attrs[i]);
|
||||
sysdev_unregister(&per_cpu(device_mce,cpu));
|
||||
cpu_clear(cpu, mce_device_initialized);
|
||||
cpumask_clear_cpu(cpu, mce_device_initialized);
|
||||
}
|
||||
|
||||
/* Make sure there are no machine checks on offlined CPUs. */
|
||||
@@ -1162,6 +1162,8 @@ static __init int mce_init_device(void)
|
||||
if (!mce_available(&boot_cpu_data))
|
||||
return -EIO;
|
||||
|
||||
alloc_cpumask_var(&mce_device_initialized, GFP_KERNEL);
|
||||
|
||||
err = mce_init_banks();
|
||||
if (err)
|
||||
return err;
|
||||
|
@@ -485,7 +485,7 @@ static __cpuinit int threshold_create_bank(unsigned int cpu, unsigned int bank)
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
if (cpu_data(cpu).cpu_core_id && shared_bank[bank]) { /* symlink */
|
||||
i = cpumask_first(&per_cpu(cpu_core_map, cpu));
|
||||
i = cpumask_first(cpu_core_mask(cpu));
|
||||
|
||||
/* first core not up yet */
|
||||
if (cpu_data(i).cpu_core_id)
|
||||
@@ -505,7 +505,7 @@ static __cpuinit int threshold_create_bank(unsigned int cpu, unsigned int bank)
|
||||
if (err)
|
||||
goto out;
|
||||
|
||||
cpumask_copy(b->cpus, &per_cpu(cpu_core_map, cpu));
|
||||
cpumask_copy(b->cpus, cpu_core_mask(cpu));
|
||||
per_cpu(threshold_banks, cpu)[bank] = b;
|
||||
goto out;
|
||||
}
|
||||
@@ -529,7 +529,7 @@ static __cpuinit int threshold_create_bank(unsigned int cpu, unsigned int bank)
|
||||
#ifndef CONFIG_SMP
|
||||
cpumask_setall(b->cpus);
|
||||
#else
|
||||
cpumask_copy(b->cpus, &per_cpu(cpu_core_map, cpu));
|
||||
cpumask_copy(b->cpus, cpu_core_mask(cpu));
|
||||
#endif
|
||||
|
||||
per_cpu(threshold_banks, cpu)[bank] = b;
|
||||
|
@@ -249,7 +249,7 @@ void cmci_rediscover(int dying)
|
||||
for_each_online_cpu (cpu) {
|
||||
if (cpu == dying)
|
||||
continue;
|
||||
if (set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)))
|
||||
if (set_cpus_allowed_ptr(current, cpumask_of(cpu)))
|
||||
continue;
|
||||
/* Recheck banks in case CPUs don't all have the same */
|
||||
if (cmci_supported(&banks))
|
||||
|
@@ -14,7 +14,7 @@ static void show_cpuinfo_core(struct seq_file *m, struct cpuinfo_x86 *c,
|
||||
if (c->x86_max_cores * smp_num_siblings > 1) {
|
||||
seq_printf(m, "physical id\t: %d\n", c->phys_proc_id);
|
||||
seq_printf(m, "siblings\t: %d\n",
|
||||
cpus_weight(per_cpu(cpu_core_map, cpu)));
|
||||
cpumask_weight(cpu_sibling_mask(cpu)));
|
||||
seq_printf(m, "core id\t\t: %d\n", c->cpu_core_id);
|
||||
seq_printf(m, "cpu cores\t: %d\n", c->booted_cores);
|
||||
seq_printf(m, "apicid\t\t: %d\n", c->apicid);
|
||||
@@ -143,9 +143,9 @@ static int show_cpuinfo(struct seq_file *m, void *v)
|
||||
static void *c_start(struct seq_file *m, loff_t *pos)
|
||||
{
|
||||
if (*pos == 0) /* just in case, cpu 0 is not the first */
|
||||
*pos = first_cpu(cpu_online_map);
|
||||
*pos = cpumask_first(cpu_online_mask);
|
||||
else
|
||||
*pos = next_cpu_nr(*pos - 1, cpu_online_map);
|
||||
*pos = cpumask_next(*pos - 1, cpu_online_mask);
|
||||
if ((*pos) < nr_cpu_ids)
|
||||
return &cpu_data(*pos);
|
||||
return NULL;
|
||||
|
@@ -12,31 +12,30 @@
|
||||
*
|
||||
* Licensed under the terms of the GNU General Public
|
||||
* License version 2. See file COPYING for details.
|
||||
*/
|
||||
|
||||
#include <linux/capability.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/miscdevice.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/mm.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/firmware.h>
|
||||
*/
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/pci.h>
|
||||
#include <linux/capability.h>
|
||||
#include <linux/miscdevice.h>
|
||||
#include <linux/firmware.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/pci_ids.h>
|
||||
#include <linux/uaccess.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/pci.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/mm.h>
|
||||
|
||||
#include <asm/msr.h>
|
||||
#include <asm/processor.h>
|
||||
#include <asm/microcode.h>
|
||||
#include <asm/processor.h>
|
||||
#include <asm/msr.h>
|
||||
|
||||
MODULE_DESCRIPTION("AMD Microcode Update Driver");
|
||||
MODULE_AUTHOR("Peter Oruba");
|
||||
@@ -72,8 +71,8 @@ struct microcode_header_amd {
|
||||
} __attribute__((packed));
|
||||
|
||||
struct microcode_amd {
|
||||
struct microcode_header_amd hdr;
|
||||
unsigned int mpb[0];
|
||||
struct microcode_header_amd hdr;
|
||||
unsigned int mpb[0];
|
||||
};
|
||||
|
||||
#define UCODE_MAX_SIZE 2048
|
||||
@@ -184,8 +183,8 @@ static int get_ucode_data(void *to, const u8 *from, size_t n)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void *get_next_ucode(const u8 *buf, unsigned int size,
|
||||
unsigned int *mc_size)
|
||||
static void *
|
||||
get_next_ucode(const u8 *buf, unsigned int size, unsigned int *mc_size)
|
||||
{
|
||||
unsigned int total_size;
|
||||
u8 section_hdr[UCODE_CONTAINER_SECTION_HDR];
|
||||
@@ -223,7 +222,6 @@ static void *get_next_ucode(const u8 *buf, unsigned int size,
|
||||
return mc;
|
||||
}
|
||||
|
||||
|
||||
static int install_equiv_cpu_table(const u8 *buf)
|
||||
{
|
||||
u8 *container_hdr[UCODE_CONTAINER_HEADER_SIZE];
|
||||
@@ -372,4 +370,3 @@ struct microcode_ops * __init init_amd_microcode(void)
|
||||
{
|
||||
return µcode_amd_ops;
|
||||
}
|
||||
|
||||
|
@@ -70,67 +70,78 @@
|
||||
* Fix sigmatch() macro to handle old CPUs with pf == 0.
|
||||
* Thanks to Stuart Swales for pointing out this bug.
|
||||
*/
|
||||
#include <linux/capability.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/smp_lock.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/miscdevice.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/mm.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/firmware.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/capability.h>
|
||||
#include <linux/miscdevice.h>
|
||||
#include <linux/firmware.h>
|
||||
#include <linux/smp_lock.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/uaccess.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/mm.h>
|
||||
|
||||
#include <asm/msr.h>
|
||||
#include <asm/uaccess.h>
|
||||
#include <asm/processor.h>
|
||||
#include <asm/microcode.h>
|
||||
#include <asm/processor.h>
|
||||
#include <asm/msr.h>
|
||||
|
||||
MODULE_DESCRIPTION("Microcode Update Driver");
|
||||
MODULE_AUTHOR("Tigran Aivazian <tigran@aivazian.fsnet.co.uk>");
|
||||
MODULE_LICENSE("GPL");
|
||||
|
||||
#define MICROCODE_VERSION "2.00"
|
||||
#define MICROCODE_VERSION "2.00"
|
||||
|
||||
static struct microcode_ops *microcode_ops;
|
||||
static struct microcode_ops *microcode_ops;
|
||||
|
||||
/* no concurrent ->write()s are allowed on /dev/cpu/microcode */
|
||||
static DEFINE_MUTEX(microcode_mutex);
|
||||
|
||||
struct ucode_cpu_info ucode_cpu_info[NR_CPUS];
|
||||
struct ucode_cpu_info ucode_cpu_info[NR_CPUS];
|
||||
EXPORT_SYMBOL_GPL(ucode_cpu_info);
|
||||
|
||||
#ifdef CONFIG_MICROCODE_OLD_INTERFACE
|
||||
struct update_for_cpu {
|
||||
const void __user *buf;
|
||||
size_t size;
|
||||
};
|
||||
|
||||
static long update_for_cpu(void *_ufc)
|
||||
{
|
||||
struct update_for_cpu *ufc = _ufc;
|
||||
int error;
|
||||
|
||||
error = microcode_ops->request_microcode_user(smp_processor_id(),
|
||||
ufc->buf, ufc->size);
|
||||
if (error < 0)
|
||||
return error;
|
||||
if (!error)
|
||||
microcode_ops->apply_microcode(smp_processor_id());
|
||||
return error;
|
||||
}
|
||||
|
||||
static int do_microcode_update(const void __user *buf, size_t size)
|
||||
{
|
||||
cpumask_t old;
|
||||
int error = 0;
|
||||
int cpu;
|
||||
|
||||
old = current->cpus_allowed;
|
||||
struct update_for_cpu ufc = { .buf = buf, .size = size };
|
||||
|
||||
for_each_online_cpu(cpu) {
|
||||
struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
|
||||
|
||||
if (!uci->valid)
|
||||
continue;
|
||||
|
||||
set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu));
|
||||
error = microcode_ops->request_microcode_user(cpu, buf, size);
|
||||
error = work_on_cpu(cpu, update_for_cpu, &ufc);
|
||||
if (error < 0)
|
||||
goto out;
|
||||
if (!error)
|
||||
microcode_ops->apply_microcode(cpu);
|
||||
break;
|
||||
}
|
||||
out:
|
||||
set_cpus_allowed_ptr(current, &old);
|
||||
return error;
|
||||
}
|
||||
|
||||
@@ -198,18 +209,33 @@ static void microcode_dev_exit(void)
|
||||
|
||||
MODULE_ALIAS_MISCDEV(MICROCODE_MINOR);
|
||||
#else
|
||||
#define microcode_dev_init() 0
|
||||
#define microcode_dev_exit() do { } while (0)
|
||||
#define microcode_dev_init() 0
|
||||
#define microcode_dev_exit() do { } while (0)
|
||||
#endif
|
||||
|
||||
/* fake device for request_firmware */
|
||||
static struct platform_device *microcode_pdev;
|
||||
static struct platform_device *microcode_pdev;
|
||||
|
||||
static long reload_for_cpu(void *unused)
|
||||
{
|
||||
struct ucode_cpu_info *uci = ucode_cpu_info + smp_processor_id();
|
||||
int err = 0;
|
||||
|
||||
mutex_lock(µcode_mutex);
|
||||
if (uci->valid) {
|
||||
err = microcode_ops->request_microcode_fw(smp_processor_id(),
|
||||
µcode_pdev->dev);
|
||||
if (!err)
|
||||
microcode_ops->apply_microcode(smp_processor_id());
|
||||
}
|
||||
mutex_unlock(µcode_mutex);
|
||||
return err;
|
||||
}
|
||||
|
||||
static ssize_t reload_store(struct sys_device *dev,
|
||||
struct sysdev_attribute *attr,
|
||||
const char *buf, size_t sz)
|
||||
{
|
||||
struct ucode_cpu_info *uci = ucode_cpu_info + dev->id;
|
||||
char *end;
|
||||
unsigned long val = simple_strtoul(buf, &end, 0);
|
||||
int err = 0;
|
||||
@@ -218,21 +244,9 @@ static ssize_t reload_store(struct sys_device *dev,
|
||||
if (end == buf)
|
||||
return -EINVAL;
|
||||
if (val == 1) {
|
||||
cpumask_t old = current->cpus_allowed;
|
||||
|
||||
get_online_cpus();
|
||||
if (cpu_online(cpu)) {
|
||||
set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu));
|
||||
mutex_lock(µcode_mutex);
|
||||
if (uci->valid) {
|
||||
err = microcode_ops->request_microcode_fw(cpu,
|
||||
µcode_pdev->dev);
|
||||
if (!err)
|
||||
microcode_ops->apply_microcode(cpu);
|
||||
}
|
||||
mutex_unlock(µcode_mutex);
|
||||
set_cpus_allowed_ptr(current, &old);
|
||||
}
|
||||
if (cpu_online(cpu))
|
||||
err = work_on_cpu(cpu, reload_for_cpu, NULL);
|
||||
put_online_cpus();
|
||||
}
|
||||
if (err)
|
||||
@@ -268,8 +282,8 @@ static struct attribute *mc_default_attrs[] = {
|
||||
};
|
||||
|
||||
static struct attribute_group mc_attr_group = {
|
||||
.attrs = mc_default_attrs,
|
||||
.name = "microcode",
|
||||
.attrs = mc_default_attrs,
|
||||
.name = "microcode",
|
||||
};
|
||||
|
||||
static void __microcode_fini_cpu(int cpu)
|
||||
@@ -328,9 +342,9 @@ static int microcode_resume_cpu(int cpu)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void microcode_update_cpu(int cpu)
|
||||
static long microcode_update_cpu(void *unused)
|
||||
{
|
||||
struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
|
||||
struct ucode_cpu_info *uci = ucode_cpu_info + smp_processor_id();
|
||||
int err = 0;
|
||||
|
||||
/*
|
||||
@@ -338,30 +352,27 @@ static void microcode_update_cpu(int cpu)
|
||||
* otherwise just request a firmware:
|
||||
*/
|
||||
if (uci->valid) {
|
||||
err = microcode_resume_cpu(cpu);
|
||||
} else {
|
||||
collect_cpu_info(cpu);
|
||||
err = microcode_resume_cpu(smp_processor_id());
|
||||
} else {
|
||||
collect_cpu_info(smp_processor_id());
|
||||
if (uci->valid && system_state == SYSTEM_RUNNING)
|
||||
err = microcode_ops->request_microcode_fw(cpu,
|
||||
err = microcode_ops->request_microcode_fw(
|
||||
smp_processor_id(),
|
||||
µcode_pdev->dev);
|
||||
}
|
||||
if (!err)
|
||||
microcode_ops->apply_microcode(cpu);
|
||||
microcode_ops->apply_microcode(smp_processor_id());
|
||||
return err;
|
||||
}
|
||||
|
||||
static void microcode_init_cpu(int cpu)
|
||||
static int microcode_init_cpu(int cpu)
|
||||
{
|
||||
cpumask_t old = current->cpus_allowed;
|
||||
|
||||
set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu));
|
||||
/* We should bind the task to the CPU */
|
||||
BUG_ON(raw_smp_processor_id() != cpu);
|
||||
|
||||
int err;
|
||||
mutex_lock(µcode_mutex);
|
||||
microcode_update_cpu(cpu);
|
||||
err = work_on_cpu(cpu, microcode_update_cpu, NULL);
|
||||
mutex_unlock(µcode_mutex);
|
||||
|
||||
set_cpus_allowed_ptr(current, &old);
|
||||
return err;
|
||||
}
|
||||
|
||||
static int mc_sysdev_add(struct sys_device *sys_dev)
|
||||
@@ -379,8 +390,11 @@ static int mc_sysdev_add(struct sys_device *sys_dev)
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
microcode_init_cpu(cpu);
|
||||
return 0;
|
||||
err = microcode_init_cpu(cpu);
|
||||
if (err)
|
||||
sysfs_remove_group(&sys_dev->kobj, &mc_attr_group);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
static int mc_sysdev_remove(struct sys_device *sys_dev)
|
||||
@@ -404,14 +418,14 @@ static int mc_sysdev_resume(struct sys_device *dev)
|
||||
return 0;
|
||||
|
||||
/* only CPU 0 will apply ucode here */
|
||||
microcode_update_cpu(0);
|
||||
microcode_update_cpu(NULL);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct sysdev_driver mc_sysdev_driver = {
|
||||
.add = mc_sysdev_add,
|
||||
.remove = mc_sysdev_remove,
|
||||
.resume = mc_sysdev_resume,
|
||||
.add = mc_sysdev_add,
|
||||
.remove = mc_sysdev_remove,
|
||||
.resume = mc_sysdev_resume,
|
||||
};
|
||||
|
||||
static __cpuinit int
|
||||
@@ -424,7 +438,9 @@ mc_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu)
|
||||
switch (action) {
|
||||
case CPU_ONLINE:
|
||||
case CPU_ONLINE_FROZEN:
|
||||
microcode_init_cpu(cpu);
|
||||
if (microcode_init_cpu(cpu))
|
||||
printk(KERN_ERR "microcode: failed to init CPU%d\n",
|
||||
cpu);
|
||||
case CPU_DOWN_FAILED:
|
||||
case CPU_DOWN_FAILED_FROZEN:
|
||||
pr_debug("microcode: CPU%d added\n", cpu);
|
||||
@@ -448,7 +464,7 @@ mc_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu)
|
||||
}
|
||||
|
||||
static struct notifier_block __refdata mc_cpu_notifier = {
|
||||
.notifier_call = mc_cpu_callback,
|
||||
.notifier_call = mc_cpu_callback,
|
||||
};
|
||||
|
||||
static int __init microcode_init(void)
|
||||
|
@@ -70,28 +70,28 @@
|
||||
* Fix sigmatch() macro to handle old CPUs with pf == 0.
|
||||
* Thanks to Stuart Swales for pointing out this bug.
|
||||
*/
|
||||
#include <linux/capability.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/smp_lock.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/miscdevice.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/mm.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/firmware.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/capability.h>
|
||||
#include <linux/miscdevice.h>
|
||||
#include <linux/firmware.h>
|
||||
#include <linux/smp_lock.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/uaccess.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/fs.h>
|
||||
#include <linux/mm.h>
|
||||
|
||||
#include <asm/msr.h>
|
||||
#include <asm/processor.h>
|
||||
#include <asm/microcode.h>
|
||||
#include <asm/processor.h>
|
||||
#include <asm/msr.h>
|
||||
|
||||
MODULE_DESCRIPTION("Microcode Update Driver");
|
||||
MODULE_AUTHOR("Tigran Aivazian <tigran@aivazian.fsnet.co.uk>");
|
||||
@@ -129,12 +129,13 @@ struct extended_sigtable {
|
||||
struct extended_signature sigs[0];
|
||||
};
|
||||
|
||||
#define DEFAULT_UCODE_DATASIZE (2000)
|
||||
#define DEFAULT_UCODE_DATASIZE (2000)
|
||||
#define MC_HEADER_SIZE (sizeof(struct microcode_header_intel))
|
||||
#define DEFAULT_UCODE_TOTALSIZE (DEFAULT_UCODE_DATASIZE + MC_HEADER_SIZE)
|
||||
#define EXT_HEADER_SIZE (sizeof(struct extended_sigtable))
|
||||
#define EXT_SIGNATURE_SIZE (sizeof(struct extended_signature))
|
||||
#define DWSIZE (sizeof(u32))
|
||||
|
||||
#define get_totalsize(mc) \
|
||||
(((struct microcode_intel *)mc)->hdr.totalsize ? \
|
||||
((struct microcode_intel *)mc)->hdr.totalsize : \
|
||||
@@ -197,30 +198,31 @@ static inline int update_match_cpu(struct cpu_signature *csig, int sig, int pf)
|
||||
}
|
||||
|
||||
static inline int
|
||||
update_match_revision(struct microcode_header_intel *mc_header, int rev)
|
||||
update_match_revision(struct microcode_header_intel *mc_header, int rev)
|
||||
{
|
||||
return (mc_header->rev <= rev) ? 0 : 1;
|
||||
}
|
||||
|
||||
static int microcode_sanity_check(void *mc)
|
||||
{
|
||||
unsigned long total_size, data_size, ext_table_size;
|
||||
struct microcode_header_intel *mc_header = mc;
|
||||
struct extended_sigtable *ext_header = NULL;
|
||||
struct extended_signature *ext_sig;
|
||||
unsigned long total_size, data_size, ext_table_size;
|
||||
int sum, orig_sum, ext_sigcount = 0, i;
|
||||
struct extended_signature *ext_sig;
|
||||
|
||||
total_size = get_totalsize(mc_header);
|
||||
data_size = get_datasize(mc_header);
|
||||
|
||||
if (data_size + MC_HEADER_SIZE > total_size) {
|
||||
printk(KERN_ERR "microcode: error! "
|
||||
"Bad data size in microcode data file\n");
|
||||
"Bad data size in microcode data file\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (mc_header->ldrver != 1 || mc_header->hdrver != 1) {
|
||||
printk(KERN_ERR "microcode: error! "
|
||||
"Unknown microcode update format\n");
|
||||
"Unknown microcode update format\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
ext_table_size = total_size - (MC_HEADER_SIZE + data_size);
|
||||
@@ -318,11 +320,15 @@ get_matching_microcode(struct cpu_signature *cpu_sig, void *mc, int rev)
|
||||
|
||||
static void apply_microcode(int cpu)
|
||||
{
|
||||
struct microcode_intel *mc_intel;
|
||||
struct ucode_cpu_info *uci;
|
||||
unsigned long flags;
|
||||
unsigned int val[2];
|
||||
int cpu_num = raw_smp_processor_id();
|
||||
struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
|
||||
struct microcode_intel *mc_intel = uci->mc;
|
||||
int cpu_num;
|
||||
|
||||
cpu_num = raw_smp_processor_id();
|
||||
uci = ucode_cpu_info + cpu;
|
||||
mc_intel = uci->mc;
|
||||
|
||||
/* We should bind the task to the CPU */
|
||||
BUG_ON(cpu_num != cpu);
|
||||
@@ -348,15 +354,17 @@ static void apply_microcode(int cpu)
|
||||
spin_unlock_irqrestore(µcode_update_lock, flags);
|
||||
if (val[1] != mc_intel->hdr.rev) {
|
||||
printk(KERN_ERR "microcode: CPU%d update from revision "
|
||||
"0x%x to 0x%x failed\n", cpu_num, uci->cpu_sig.rev, val[1]);
|
||||
"0x%x to 0x%x failed\n",
|
||||
cpu_num, uci->cpu_sig.rev, val[1]);
|
||||
return;
|
||||
}
|
||||
printk(KERN_INFO "microcode: CPU%d updated from revision "
|
||||
"0x%x to 0x%x, date = %04x-%02x-%02x \n",
|
||||
"0x%x to 0x%x, date = %04x-%02x-%02x \n",
|
||||
cpu_num, uci->cpu_sig.rev, val[1],
|
||||
mc_intel->hdr.date & 0xffff,
|
||||
mc_intel->hdr.date >> 24,
|
||||
(mc_intel->hdr.date >> 16) & 0xff);
|
||||
|
||||
uci->cpu_sig.rev = val[1];
|
||||
}
|
||||
|
||||
@@ -404,18 +412,23 @@ static int generic_load_microcode(int cpu, void *data, size_t size,
|
||||
leftover -= mc_size;
|
||||
}
|
||||
|
||||
if (new_mc) {
|
||||
if (!leftover) {
|
||||
if (uci->mc)
|
||||
vfree(uci->mc);
|
||||
uci->mc = (struct microcode_intel *)new_mc;
|
||||
pr_debug("microcode: CPU%d found a matching microcode update with"
|
||||
" version 0x%x (current=0x%x)\n",
|
||||
cpu, new_rev, uci->cpu_sig.rev);
|
||||
} else
|
||||
vfree(new_mc);
|
||||
if (!new_mc)
|
||||
goto out;
|
||||
|
||||
if (leftover) {
|
||||
vfree(new_mc);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (uci->mc)
|
||||
vfree(uci->mc);
|
||||
uci->mc = (struct microcode_intel *)new_mc;
|
||||
|
||||
pr_debug("microcode: CPU%d found a matching microcode update with"
|
||||
" version 0x%x (current=0x%x)\n",
|
||||
cpu, new_rev, uci->cpu_sig.rev);
|
||||
|
||||
out:
|
||||
return (int)leftover;
|
||||
}
|
||||
|
||||
|
@@ -325,7 +325,7 @@ void stop_this_cpu(void *dummy)
|
||||
/*
|
||||
* Remove this CPU:
|
||||
*/
|
||||
cpu_clear(smp_processor_id(), cpu_online_map);
|
||||
set_cpu_online(smp_processor_id(), false);
|
||||
disable_local_APIC();
|
||||
|
||||
for (;;) {
|
||||
@@ -475,12 +475,13 @@ static int __cpuinit check_c1e_idle(const struct cpuinfo_x86 *c)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static cpumask_t c1e_mask = CPU_MASK_NONE;
|
||||
static cpumask_var_t c1e_mask;
|
||||
static int c1e_detected;
|
||||
|
||||
void c1e_remove_cpu(int cpu)
|
||||
{
|
||||
cpu_clear(cpu, c1e_mask);
|
||||
if (c1e_mask != NULL)
|
||||
cpumask_clear_cpu(cpu, c1e_mask);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -509,8 +510,8 @@ static void c1e_idle(void)
|
||||
if (c1e_detected) {
|
||||
int cpu = smp_processor_id();
|
||||
|
||||
if (!cpu_isset(cpu, c1e_mask)) {
|
||||
cpu_set(cpu, c1e_mask);
|
||||
if (!cpumask_test_cpu(cpu, c1e_mask)) {
|
||||
cpumask_set_cpu(cpu, c1e_mask);
|
||||
/*
|
||||
* Force broadcast so ACPI can not interfere. Needs
|
||||
* to run with interrupts enabled as it uses
|
||||
@@ -562,6 +563,15 @@ void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c)
|
||||
pm_idle = default_idle;
|
||||
}
|
||||
|
||||
void __init init_c1e_mask(void)
|
||||
{
|
||||
/* If we're using c1e_idle, we need to allocate c1e_mask. */
|
||||
if (pm_idle == c1e_idle) {
|
||||
alloc_cpumask_var(&c1e_mask, GFP_KERNEL);
|
||||
cpumask_clear(c1e_mask);
|
||||
}
|
||||
}
|
||||
|
||||
static int __init idle_setup(char *str)
|
||||
{
|
||||
if (!str)
|
||||
|
@@ -101,11 +101,11 @@ EXPORT_SYMBOL(smp_num_siblings);
|
||||
DEFINE_PER_CPU(u16, cpu_llc_id) = BAD_APICID;
|
||||
|
||||
/* representing HT siblings of each logical CPU */
|
||||
DEFINE_PER_CPU(cpumask_t, cpu_sibling_map);
|
||||
DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map);
|
||||
EXPORT_PER_CPU_SYMBOL(cpu_sibling_map);
|
||||
|
||||
/* representing HT and core siblings of each logical CPU */
|
||||
DEFINE_PER_CPU(cpumask_t, cpu_core_map);
|
||||
DEFINE_PER_CPU(cpumask_var_t, cpu_core_map);
|
||||
EXPORT_PER_CPU_SYMBOL(cpu_core_map);
|
||||
|
||||
/* Per CPU bogomips and other parameters */
|
||||
@@ -115,11 +115,6 @@ EXPORT_PER_CPU_SYMBOL(cpu_info);
|
||||
atomic_t init_deasserted;
|
||||
|
||||
#if defined(CONFIG_NUMA) && defined(CONFIG_X86_32)
|
||||
|
||||
/* which logical CPUs are on which nodes */
|
||||
cpumask_t node_to_cpumask_map[MAX_NUMNODES] __read_mostly =
|
||||
{ [0 ... MAX_NUMNODES-1] = CPU_MASK_NONE };
|
||||
EXPORT_SYMBOL(node_to_cpumask_map);
|
||||
/* which node each logical CPU is on */
|
||||
int cpu_to_node_map[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = 0 };
|
||||
EXPORT_SYMBOL(cpu_to_node_map);
|
||||
@@ -128,7 +123,7 @@ EXPORT_SYMBOL(cpu_to_node_map);
|
||||
static void map_cpu_to_node(int cpu, int node)
|
||||
{
|
||||
printk(KERN_INFO "Mapping cpu %d to node %d\n", cpu, node);
|
||||
cpumask_set_cpu(cpu, &node_to_cpumask_map[node]);
|
||||
cpumask_set_cpu(cpu, node_to_cpumask_map[node]);
|
||||
cpu_to_node_map[cpu] = node;
|
||||
}
|
||||
|
||||
@@ -139,7 +134,7 @@ static void unmap_cpu_to_node(int cpu)
|
||||
|
||||
printk(KERN_INFO "Unmapping cpu %d from all nodes\n", cpu);
|
||||
for (node = 0; node < MAX_NUMNODES; node++)
|
||||
cpumask_clear_cpu(cpu, &node_to_cpumask_map[node]);
|
||||
cpumask_clear_cpu(cpu, node_to_cpumask_map[node]);
|
||||
cpu_to_node_map[cpu] = 0;
|
||||
}
|
||||
#else /* !(CONFIG_NUMA && CONFIG_X86_32) */
|
||||
@@ -301,7 +296,7 @@ notrace static void __cpuinit start_secondary(void *unused)
|
||||
__flush_tlb_all();
|
||||
#endif
|
||||
|
||||
/* This must be done before setting cpu_online_map */
|
||||
/* This must be done before setting cpu_online_mask */
|
||||
set_cpu_sibling_map(raw_smp_processor_id());
|
||||
wmb();
|
||||
|
||||
@@ -334,6 +329,23 @@ notrace static void __cpuinit start_secondary(void *unused)
|
||||
cpu_idle();
|
||||
}
|
||||
|
||||
#ifdef CONFIG_CPUMASK_OFFSTACK
|
||||
/* In this case, llc_shared_map is a pointer to a cpumask. */
|
||||
static inline void copy_cpuinfo_x86(struct cpuinfo_x86 *dst,
|
||||
const struct cpuinfo_x86 *src)
|
||||
{
|
||||
struct cpumask *llc = dst->llc_shared_map;
|
||||
*dst = *src;
|
||||
dst->llc_shared_map = llc;
|
||||
}
|
||||
#else
|
||||
static inline void copy_cpuinfo_x86(struct cpuinfo_x86 *dst,
|
||||
const struct cpuinfo_x86 *src)
|
||||
{
|
||||
*dst = *src;
|
||||
}
|
||||
#endif /* CONFIG_CPUMASK_OFFSTACK */
|
||||
|
||||
/*
|
||||
* The bootstrap kernel entry code has set these up. Save them for
|
||||
* a given CPU
|
||||
@@ -343,7 +355,7 @@ void __cpuinit smp_store_cpu_info(int id)
|
||||
{
|
||||
struct cpuinfo_x86 *c = &cpu_data(id);
|
||||
|
||||
*c = boot_cpu_data;
|
||||
copy_cpuinfo_x86(c, &boot_cpu_data);
|
||||
c->cpu_index = id;
|
||||
if (id != 0)
|
||||
identify_secondary_cpu(c);
|
||||
@@ -367,15 +379,15 @@ void __cpuinit set_cpu_sibling_map(int cpu)
|
||||
cpumask_set_cpu(cpu, cpu_sibling_mask(i));
|
||||
cpumask_set_cpu(i, cpu_core_mask(cpu));
|
||||
cpumask_set_cpu(cpu, cpu_core_mask(i));
|
||||
cpumask_set_cpu(i, &c->llc_shared_map);
|
||||
cpumask_set_cpu(cpu, &o->llc_shared_map);
|
||||
cpumask_set_cpu(i, c->llc_shared_map);
|
||||
cpumask_set_cpu(cpu, o->llc_shared_map);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
cpumask_set_cpu(cpu, cpu_sibling_mask(cpu));
|
||||
}
|
||||
|
||||
cpumask_set_cpu(cpu, &c->llc_shared_map);
|
||||
cpumask_set_cpu(cpu, c->llc_shared_map);
|
||||
|
||||
if (current_cpu_data.x86_max_cores == 1) {
|
||||
cpumask_copy(cpu_core_mask(cpu), cpu_sibling_mask(cpu));
|
||||
@@ -386,8 +398,8 @@ void __cpuinit set_cpu_sibling_map(int cpu)
|
||||
for_each_cpu(i, cpu_sibling_setup_mask) {
|
||||
if (per_cpu(cpu_llc_id, cpu) != BAD_APICID &&
|
||||
per_cpu(cpu_llc_id, cpu) == per_cpu(cpu_llc_id, i)) {
|
||||
cpumask_set_cpu(i, &c->llc_shared_map);
|
||||
cpumask_set_cpu(cpu, &cpu_data(i).llc_shared_map);
|
||||
cpumask_set_cpu(i, c->llc_shared_map);
|
||||
cpumask_set_cpu(cpu, cpu_data(i).llc_shared_map);
|
||||
}
|
||||
if (c->phys_proc_id == cpu_data(i).phys_proc_id) {
|
||||
cpumask_set_cpu(i, cpu_core_mask(cpu));
|
||||
@@ -425,12 +437,7 @@ const struct cpumask *cpu_coregroup_mask(int cpu)
|
||||
if (sched_mc_power_savings || sched_smt_power_savings)
|
||||
return cpu_core_mask(cpu);
|
||||
else
|
||||
return &c->llc_shared_map;
|
||||
}
|
||||
|
||||
cpumask_t cpu_coregroup_map(int cpu)
|
||||
{
|
||||
return *cpu_coregroup_mask(cpu);
|
||||
return c->llc_shared_map;
|
||||
}
|
||||
|
||||
static void impress_friends(void)
|
||||
@@ -897,9 +904,8 @@ int __cpuinit native_cpu_up(unsigned int cpu)
|
||||
*/
|
||||
static __init void disable_smp(void)
|
||||
{
|
||||
/* use the read/write pointers to the present and possible maps */
|
||||
cpumask_copy(&cpu_present_map, cpumask_of(0));
|
||||
cpumask_copy(&cpu_possible_map, cpumask_of(0));
|
||||
init_cpu_present(cpumask_of(0));
|
||||
init_cpu_possible(cpumask_of(0));
|
||||
smpboot_clear_io_apic_irqs();
|
||||
|
||||
if (smp_found_config)
|
||||
@@ -1031,6 +1037,8 @@ static void __init smp_cpu_index_default(void)
|
||||
*/
|
||||
void __init native_smp_prepare_cpus(unsigned int max_cpus)
|
||||
{
|
||||
unsigned int i;
|
||||
|
||||
preempt_disable();
|
||||
smp_cpu_index_default();
|
||||
current_cpu_data = boot_cpu_data;
|
||||
@@ -1044,6 +1052,14 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus)
|
||||
boot_cpu_logical_apicid = logical_smp_processor_id();
|
||||
#endif
|
||||
current_thread_info()->cpu = 0; /* needed? */
|
||||
for_each_possible_cpu(i) {
|
||||
alloc_cpumask_var(&per_cpu(cpu_sibling_map, i), GFP_KERNEL);
|
||||
alloc_cpumask_var(&per_cpu(cpu_core_map, i), GFP_KERNEL);
|
||||
alloc_cpumask_var(&cpu_data(i).llc_shared_map, GFP_KERNEL);
|
||||
cpumask_clear(per_cpu(cpu_core_map, i));
|
||||
cpumask_clear(per_cpu(cpu_sibling_map, i));
|
||||
cpumask_clear(cpu_data(i).llc_shared_map);
|
||||
}
|
||||
set_cpu_sibling_map(0);
|
||||
|
||||
enable_IR_x2apic();
|
||||
@@ -1132,11 +1148,11 @@ early_param("possible_cpus", _setup_possible_cpus);
|
||||
|
||||
|
||||
/*
|
||||
* cpu_possible_map should be static, it cannot change as cpu's
|
||||
* cpu_possible_mask should be static, it cannot change as cpu's
|
||||
* are onlined, or offlined. The reason is per-cpu data-structures
|
||||
* are allocated by some modules at init time, and dont expect to
|
||||
* do this dynamically on cpu arrival/departure.
|
||||
* cpu_present_map on the other hand can change dynamically.
|
||||
* cpu_present_mask on the other hand can change dynamically.
|
||||
* In case when cpu_hotplug is not compiled, then we resort to current
|
||||
* behaviour, which is cpu_possible == cpu_present.
|
||||
* - Ashok Raj
|
||||
|
@@ -275,6 +275,8 @@ const struct cpumask *uv_flush_send_and_wait(int cpu, int this_blade,
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static DEFINE_PER_CPU(cpumask_var_t, uv_flush_tlb_mask);
|
||||
|
||||
/**
|
||||
* uv_flush_tlb_others - globally purge translation cache of a virtual
|
||||
* address or all TLB's
|
||||
@@ -304,8 +306,7 @@ const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask,
|
||||
struct mm_struct *mm,
|
||||
unsigned long va, unsigned int cpu)
|
||||
{
|
||||
static DEFINE_PER_CPU(cpumask_t, flush_tlb_mask);
|
||||
struct cpumask *flush_mask = &__get_cpu_var(flush_tlb_mask);
|
||||
struct cpumask *flush_mask = __get_cpu_var(uv_flush_tlb_mask);
|
||||
int i;
|
||||
int bit;
|
||||
int blade;
|
||||
@@ -755,6 +756,10 @@ static int __init uv_bau_init(void)
|
||||
if (!is_uv_system())
|
||||
return 0;
|
||||
|
||||
for_each_possible_cpu(cur_cpu)
|
||||
alloc_cpumask_var_node(&per_cpu(uv_flush_tlb_mask, cur_cpu),
|
||||
GFP_KERNEL, cpu_to_node(cur_cpu));
|
||||
|
||||
uv_bau_retry_limit = 1;
|
||||
uv_nshift = uv_hub_info->n_val;
|
||||
uv_mmask = (1UL << uv_hub_info->n_val) - 1;
|
||||
|
Reference in New Issue
Block a user