[IA64] remove remaining __FUNCTION__ occurrences
__FUNCTION__ is gcc-specific, use __func__ Long lines have been kept where they exist, some small spacing changes have been done. Signed-off-by: Harvey Harrison <harvey.harrison@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Tony Luck <tony.luck@intel.com>
This commit is contained in:

committed by
Tony Luck

parent
2d9b06c72a
commit
d4ed80841a
@@ -413,8 +413,8 @@ ia64_log_get(int sal_info_type, u8 **buffer, int irq_safe)
|
||||
IA64_LOG_INDEX_INC(sal_info_type);
|
||||
IA64_LOG_UNLOCK(sal_info_type);
|
||||
if (irq_safe) {
|
||||
IA64_MCA_DEBUG("%s: SAL error record type %d retrieved. "
|
||||
"Record length = %ld\n", __FUNCTION__, sal_info_type, total_len);
|
||||
IA64_MCA_DEBUG("%s: SAL error record type %d retrieved. Record length = %ld\n",
|
||||
__func__, sal_info_type, total_len);
|
||||
}
|
||||
*buffer = (u8 *) log_buffer;
|
||||
return total_len;
|
||||
@@ -518,7 +518,7 @@ ia64_mca_cpe_int_handler (int cpe_irq, void *arg)
|
||||
static DEFINE_SPINLOCK(cpe_history_lock);
|
||||
|
||||
IA64_MCA_DEBUG("%s: received interrupt vector = %#x on CPU %d\n",
|
||||
__FUNCTION__, cpe_irq, smp_processor_id());
|
||||
__func__, cpe_irq, smp_processor_id());
|
||||
|
||||
/* SAL spec states this should run w/ interrupts enabled */
|
||||
local_irq_enable();
|
||||
@@ -594,7 +594,7 @@ ia64_mca_register_cpev (int cpev)
|
||||
}
|
||||
|
||||
IA64_MCA_DEBUG("%s: corrected platform error "
|
||||
"vector %#x registered\n", __FUNCTION__, cpev);
|
||||
"vector %#x registered\n", __func__, cpev);
|
||||
}
|
||||
#endif /* CONFIG_ACPI */
|
||||
|
||||
@@ -621,12 +621,11 @@ ia64_mca_cmc_vector_setup (void)
|
||||
cmcv.cmcv_vector = IA64_CMC_VECTOR;
|
||||
ia64_setreg(_IA64_REG_CR_CMCV, cmcv.cmcv_regval);
|
||||
|
||||
IA64_MCA_DEBUG("%s: CPU %d corrected "
|
||||
"machine check vector %#x registered.\n",
|
||||
__FUNCTION__, smp_processor_id(), IA64_CMC_VECTOR);
|
||||
IA64_MCA_DEBUG("%s: CPU %d corrected machine check vector %#x registered.\n",
|
||||
__func__, smp_processor_id(), IA64_CMC_VECTOR);
|
||||
|
||||
IA64_MCA_DEBUG("%s: CPU %d CMCV = %#016lx\n",
|
||||
__FUNCTION__, smp_processor_id(), ia64_getreg(_IA64_REG_CR_CMCV));
|
||||
__func__, smp_processor_id(), ia64_getreg(_IA64_REG_CR_CMCV));
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -651,9 +650,8 @@ ia64_mca_cmc_vector_disable (void *dummy)
|
||||
cmcv.cmcv_mask = 1; /* Mask/disable interrupt */
|
||||
ia64_setreg(_IA64_REG_CR_CMCV, cmcv.cmcv_regval);
|
||||
|
||||
IA64_MCA_DEBUG("%s: CPU %d corrected "
|
||||
"machine check vector %#x disabled.\n",
|
||||
__FUNCTION__, smp_processor_id(), cmcv.cmcv_vector);
|
||||
IA64_MCA_DEBUG("%s: CPU %d corrected machine check vector %#x disabled.\n",
|
||||
__func__, smp_processor_id(), cmcv.cmcv_vector);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -678,9 +676,8 @@ ia64_mca_cmc_vector_enable (void *dummy)
|
||||
cmcv.cmcv_mask = 0; /* Unmask/enable interrupt */
|
||||
ia64_setreg(_IA64_REG_CR_CMCV, cmcv.cmcv_regval);
|
||||
|
||||
IA64_MCA_DEBUG("%s: CPU %d corrected "
|
||||
"machine check vector %#x enabled.\n",
|
||||
__FUNCTION__, smp_processor_id(), cmcv.cmcv_vector);
|
||||
IA64_MCA_DEBUG("%s: CPU %d corrected machine check vector %#x enabled.\n",
|
||||
__func__, smp_processor_id(), cmcv.cmcv_vector);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -767,7 +764,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg)
|
||||
local_irq_save(flags);
|
||||
if (notify_die(DIE_MCA_RENDZVOUS_ENTER, "MCA", get_irq_regs(),
|
||||
(long)&nd, 0, 0) == NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
|
||||
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_DONE;
|
||||
/* Register with the SAL monarch that the slave has
|
||||
@@ -777,7 +774,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg)
|
||||
|
||||
if (notify_die(DIE_MCA_RENDZVOUS_PROCESS, "MCA", get_irq_regs(),
|
||||
(long)&nd, 0, 0) == NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
|
||||
/* Wait for the monarch cpu to exit. */
|
||||
while (monarch_cpu != -1)
|
||||
@@ -785,7 +782,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg)
|
||||
|
||||
if (notify_die(DIE_MCA_RENDZVOUS_LEAVE, "MCA", get_irq_regs(),
|
||||
(long)&nd, 0, 0) == NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
|
||||
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_NOTDONE;
|
||||
/* Enable all interrupts */
|
||||
@@ -1230,7 +1227,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
|
||||
if (notify_die(DIE_MCA_MONARCH_ENTER, "MCA", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
|
||||
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_CONCURRENT_MCA;
|
||||
if (sos->monarch) {
|
||||
@@ -1246,7 +1243,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
ia64_mca_wakeup_all();
|
||||
if (notify_die(DIE_MCA_MONARCH_PROCESS, "MCA", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
} else {
|
||||
while (cpu_isset(cpu, mca_cpu))
|
||||
cpu_relax(); /* spin until monarch wakes us */
|
||||
@@ -1276,7 +1273,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
}
|
||||
if (notify_die(DIE_MCA_MONARCH_LEAVE, "MCA", regs, (long)&nd, 0, recover)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
|
||||
|
||||
if (atomic_dec_return(&mca_count) > 0) {
|
||||
@@ -1328,7 +1325,7 @@ ia64_mca_cmc_int_handler(int cmc_irq, void *arg)
|
||||
static DEFINE_SPINLOCK(cmc_history_lock);
|
||||
|
||||
IA64_MCA_DEBUG("%s: received interrupt vector = %#x on CPU %d\n",
|
||||
__FUNCTION__, cmc_irq, smp_processor_id());
|
||||
__func__, cmc_irq, smp_processor_id());
|
||||
|
||||
/* SAL spec states this should run w/ interrupts enabled */
|
||||
local_irq_enable();
|
||||
@@ -1614,7 +1611,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
*/
|
||||
if (!sos->monarch && atomic_add_return(1, &slaves) == num_online_cpus()) {
|
||||
mprintk(KERN_WARNING "%s: Promoting cpu %d to monarch.\n",
|
||||
__FUNCTION__, cpu);
|
||||
__func__, cpu);
|
||||
atomic_dec(&slaves);
|
||||
sos->monarch = 1;
|
||||
}
|
||||
@@ -1626,7 +1623,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
*/
|
||||
if (sos->monarch && atomic_add_return(1, &monarchs) > 1) {
|
||||
mprintk(KERN_WARNING "%s: Demoting cpu %d to slave.\n",
|
||||
__FUNCTION__, cpu);
|
||||
__func__, cpu);
|
||||
atomic_dec(&monarchs);
|
||||
sos->monarch = 0;
|
||||
}
|
||||
@@ -1637,15 +1634,15 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
cpu_relax(); /* spin until monarch enters */
|
||||
if (notify_die(DIE_INIT_SLAVE_ENTER, "INIT", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
if (notify_die(DIE_INIT_SLAVE_PROCESS, "INIT", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
while (monarch_cpu != -1)
|
||||
cpu_relax(); /* spin until monarch leaves */
|
||||
if (notify_die(DIE_INIT_SLAVE_LEAVE, "INIT", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
mprintk("Slave on cpu %d returning to normal service.\n", cpu);
|
||||
set_curr_task(cpu, previous_current);
|
||||
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_NOTDONE;
|
||||
@@ -1656,7 +1653,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
monarch_cpu = cpu;
|
||||
if (notify_die(DIE_INIT_MONARCH_ENTER, "INIT", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
|
||||
/*
|
||||
* Wait for a bit. On some machines (e.g., HP's zx2000 and zx6000, INIT can be
|
||||
@@ -1673,10 +1670,10 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
*/
|
||||
if (notify_die(DIE_INIT_MONARCH_PROCESS, "INIT", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
if (notify_die(DIE_INIT_MONARCH_LEAVE, "INIT", regs, (long)&nd, 0, 0)
|
||||
== NOTIFY_STOP)
|
||||
ia64_mca_spin(__FUNCTION__);
|
||||
ia64_mca_spin(__func__);
|
||||
mprintk("\nINIT dump complete. Monarch on cpu %d returning to normal service.\n", cpu);
|
||||
atomic_dec(&monarchs);
|
||||
set_curr_task(cpu, previous_current);
|
||||
@@ -1884,7 +1881,7 @@ ia64_mca_init(void)
|
||||
.priority = 0/* we need to notified last */
|
||||
};
|
||||
|
||||
IA64_MCA_DEBUG("%s: begin\n", __FUNCTION__);
|
||||
IA64_MCA_DEBUG("%s: begin\n", __func__);
|
||||
|
||||
/* Clear the Rendez checkin flag for all cpus */
|
||||
for(i = 0 ; i < NR_CPUS; i++)
|
||||
@@ -1928,7 +1925,7 @@ ia64_mca_init(void)
|
||||
return;
|
||||
}
|
||||
|
||||
IA64_MCA_DEBUG("%s: registered MCA rendezvous spinloop and wakeup mech.\n", __FUNCTION__);
|
||||
IA64_MCA_DEBUG("%s: registered MCA rendezvous spinloop and wakeup mech.\n", __func__);
|
||||
|
||||
ia64_mc_info.imi_mca_handler = ia64_tpa(mca_hldlr_ptr->fp);
|
||||
/*
|
||||
@@ -1949,7 +1946,7 @@ ia64_mca_init(void)
|
||||
return;
|
||||
}
|
||||
|
||||
IA64_MCA_DEBUG("%s: registered OS MCA handler with SAL at 0x%lx, gp = 0x%lx\n", __FUNCTION__,
|
||||
IA64_MCA_DEBUG("%s: registered OS MCA handler with SAL at 0x%lx, gp = 0x%lx\n", __func__,
|
||||
ia64_mc_info.imi_mca_handler, ia64_tpa(mca_hldlr_ptr->gp));
|
||||
|
||||
/*
|
||||
@@ -1961,7 +1958,7 @@ ia64_mca_init(void)
|
||||
ia64_mc_info.imi_slave_init_handler = ia64_tpa(init_hldlr_ptr_slave->fp);
|
||||
ia64_mc_info.imi_slave_init_handler_size = 0;
|
||||
|
||||
IA64_MCA_DEBUG("%s: OS INIT handler at %lx\n", __FUNCTION__,
|
||||
IA64_MCA_DEBUG("%s: OS INIT handler at %lx\n", __func__,
|
||||
ia64_mc_info.imi_monarch_init_handler);
|
||||
|
||||
/* Register the os init handler with SAL */
|
||||
@@ -1982,7 +1979,7 @@ ia64_mca_init(void)
|
||||
return;
|
||||
}
|
||||
|
||||
IA64_MCA_DEBUG("%s: registered OS INIT handler with SAL\n", __FUNCTION__);
|
||||
IA64_MCA_DEBUG("%s: registered OS INIT handler with SAL\n", __func__);
|
||||
|
||||
/*
|
||||
* Configure the CMCI/P vector and handler. Interrupts for CMC are
|
||||
@@ -2042,7 +2039,7 @@ ia64_mca_late_init(void)
|
||||
cmc_polling_enabled = 0;
|
||||
schedule_work(&cmc_enable_work);
|
||||
|
||||
IA64_MCA_DEBUG("%s: CMCI/P setup and enabled.\n", __FUNCTION__);
|
||||
IA64_MCA_DEBUG("%s: CMCI/P setup and enabled.\n", __func__);
|
||||
|
||||
#ifdef CONFIG_ACPI
|
||||
/* Setup the CPEI/P vector and handler */
|
||||
@@ -2065,17 +2062,17 @@ ia64_mca_late_init(void)
|
||||
ia64_cpe_irq = irq;
|
||||
ia64_mca_register_cpev(cpe_vector);
|
||||
IA64_MCA_DEBUG("%s: CPEI/P setup and enabled.\n",
|
||||
__FUNCTION__);
|
||||
__func__);
|
||||
return 0;
|
||||
}
|
||||
printk(KERN_ERR "%s: Failed to find irq for CPE "
|
||||
"interrupt handler, vector %d\n",
|
||||
__FUNCTION__, cpe_vector);
|
||||
__func__, cpe_vector);
|
||||
}
|
||||
/* If platform doesn't support CPEI, get the timer going. */
|
||||
if (cpe_poll_enabled) {
|
||||
ia64_mca_cpe_poll(0UL);
|
||||
IA64_MCA_DEBUG("%s: CPEP setup and enabled.\n", __FUNCTION__);
|
||||
IA64_MCA_DEBUG("%s: CPEP setup and enabled.\n", __func__);
|
||||
}
|
||||
}
|
||||
#endif
|
||||
|
Reference in New Issue
Block a user