ARM: convert printk(KERN_* to pr_*
Convert many (but not all) printk(KERN_* to pr_* to simplify the code. We take the opportunity to join some printk lines together so we don't split the message across several lines, and we also add a few levels to some messages which were previously missing them. Tested-by: Andrew Lunn <andrew@lunn.ch> Tested-by: Felipe Balbi <balbi@ti.com> Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
Este cometimento está contido em:
@@ -113,7 +113,7 @@ static int safe_usermode(int new_usermode, bool warn)
|
||||
new_usermode |= UM_FIXUP;
|
||||
|
||||
if (warn)
|
||||
printk(KERN_WARNING "alignment: ignoring faults is unsafe on this CPU. Defaulting to fixup mode.\n");
|
||||
pr_warn("alignment: ignoring faults is unsafe on this CPU. Defaulting to fixup mode.\n");
|
||||
}
|
||||
|
||||
return new_usermode;
|
||||
@@ -523,7 +523,7 @@ do_alignment_ldmstm(unsigned long addr, unsigned long instr, struct pt_regs *reg
|
||||
* processor for us.
|
||||
*/
|
||||
if (addr != eaddr) {
|
||||
printk(KERN_ERR "LDMSTM: PC = %08lx, instr = %08lx, "
|
||||
pr_err("LDMSTM: PC = %08lx, instr = %08lx, "
|
||||
"addr = %08lx, eaddr = %08lx\n",
|
||||
instruction_pointer(regs), instr, addr, eaddr);
|
||||
show_regs(regs);
|
||||
@@ -567,7 +567,7 @@ fault:
|
||||
return TYPE_FAULT;
|
||||
|
||||
bad:
|
||||
printk(KERN_ERR "Alignment trap: not handling ldm with s-bit set\n");
|
||||
pr_err("Alignment trap: not handling ldm with s-bit set\n");
|
||||
return TYPE_ERROR;
|
||||
}
|
||||
|
||||
@@ -899,13 +899,13 @@ do_alignment(unsigned long addr, unsigned int fsr, struct pt_regs *regs)
|
||||
return 0;
|
||||
|
||||
swp:
|
||||
printk(KERN_ERR "Alignment trap: not handling swp instruction\n");
|
||||
pr_err("Alignment trap: not handling swp instruction\n");
|
||||
|
||||
bad:
|
||||
/*
|
||||
* Oops, we didn't handle the instruction.
|
||||
*/
|
||||
printk(KERN_ERR "Alignment trap: not handling instruction "
|
||||
pr_err("Alignment trap: not handling instruction "
|
||||
"%0*lx at [<%08lx>]\n",
|
||||
isize << 1,
|
||||
isize == 2 ? tinstr : instr, instrptr);
|
||||
|
@@ -313,7 +313,7 @@ static void __init disable_l2_prefetch(void)
|
||||
*/
|
||||
u = read_extra_features();
|
||||
if (!(u & 0x01000000)) {
|
||||
printk(KERN_INFO "Feroceon L2: Disabling L2 prefetch.\n");
|
||||
pr_info("Feroceon L2: Disabling L2 prefetch.\n");
|
||||
write_extra_features(u | 0x01000000);
|
||||
}
|
||||
}
|
||||
@@ -326,7 +326,7 @@ static void __init enable_l2(void)
|
||||
if (!(u & 0x00400000)) {
|
||||
int i, d;
|
||||
|
||||
printk(KERN_INFO "Feroceon L2: Enabling L2\n");
|
||||
pr_info("Feroceon L2: Enabling L2\n");
|
||||
|
||||
d = flush_and_disable_dcache();
|
||||
i = invalidate_and_disable_icache();
|
||||
@@ -353,7 +353,7 @@ void __init feroceon_l2_init(int __l2_wt_override)
|
||||
|
||||
enable_l2();
|
||||
|
||||
printk(KERN_INFO "Feroceon L2: Cache support initialised%s.\n",
|
||||
pr_info("Feroceon L2: Cache support initialised%s.\n",
|
||||
l2_wt_override ? ", in WT override mode" : "");
|
||||
}
|
||||
#ifdef CONFIG_OF
|
||||
|
@@ -185,7 +185,7 @@ static void enable_extra_feature(unsigned int features)
|
||||
u &= ~0x01000000;
|
||||
else
|
||||
u |= 0x01000000;
|
||||
printk(KERN_INFO "Tauros2: %s L2 prefetch.\n",
|
||||
pr_info("Tauros2: %s L2 prefetch.\n",
|
||||
(features & CACHE_TAUROS2_PREFETCH_ON)
|
||||
? "Enabling" : "Disabling");
|
||||
|
||||
@@ -193,7 +193,7 @@ static void enable_extra_feature(unsigned int features)
|
||||
u |= 0x00100000;
|
||||
else
|
||||
u &= ~0x00100000;
|
||||
printk(KERN_INFO "Tauros2: %s line fill burt8.\n",
|
||||
pr_info("Tauros2: %s line fill burt8.\n",
|
||||
(features & CACHE_TAUROS2_LINEFILL_BURST8)
|
||||
? "Enabling" : "Disabling");
|
||||
|
||||
@@ -216,7 +216,7 @@ static void __init tauros2_internal_init(unsigned int features)
|
||||
*/
|
||||
feat = read_extra_features();
|
||||
if (!(feat & 0x00400000)) {
|
||||
printk(KERN_INFO "Tauros2: Enabling L2 cache.\n");
|
||||
pr_info("Tauros2: Enabling L2 cache.\n");
|
||||
write_extra_features(feat | 0x00400000);
|
||||
}
|
||||
|
||||
@@ -253,7 +253,7 @@ static void __init tauros2_internal_init(unsigned int features)
|
||||
*/
|
||||
actlr = read_actlr();
|
||||
if (!(actlr & 0x00000002)) {
|
||||
printk(KERN_INFO "Tauros2: Enabling L2 cache.\n");
|
||||
pr_info("Tauros2: Enabling L2 cache.\n");
|
||||
write_actlr(actlr | 0x00000002);
|
||||
}
|
||||
|
||||
@@ -262,11 +262,11 @@ static void __init tauros2_internal_init(unsigned int features)
|
||||
#endif
|
||||
|
||||
if (mode == NULL) {
|
||||
printk(KERN_CRIT "Tauros2: Unable to detect CPU mode.\n");
|
||||
pr_crit("Tauros2: Unable to detect CPU mode.\n");
|
||||
return;
|
||||
}
|
||||
|
||||
printk(KERN_INFO "Tauros2: L2 cache support initialised "
|
||||
pr_info("Tauros2: L2 cache support initialised "
|
||||
"in %s mode.\n", mode);
|
||||
}
|
||||
|
||||
|
@@ -235,7 +235,7 @@ void __init check_writebuffer_bugs(void)
|
||||
const char *reason;
|
||||
unsigned long v = 1;
|
||||
|
||||
printk(KERN_INFO "CPU: Testing write buffer coherency: ");
|
||||
pr_info("CPU: Testing write buffer coherency: ");
|
||||
|
||||
page = alloc_page(GFP_KERNEL);
|
||||
if (page) {
|
||||
@@ -261,9 +261,9 @@ void __init check_writebuffer_bugs(void)
|
||||
}
|
||||
|
||||
if (v) {
|
||||
printk("failed, %s\n", reason);
|
||||
pr_cont("failed, %s\n", reason);
|
||||
shared_pte_mask = L_PTE_MT_UNCACHED;
|
||||
} else {
|
||||
printk("ok\n");
|
||||
pr_cont("ok\n");
|
||||
}
|
||||
}
|
||||
|
@@ -63,9 +63,9 @@ void show_pte(struct mm_struct *mm, unsigned long addr)
|
||||
if (!mm)
|
||||
mm = &init_mm;
|
||||
|
||||
printk(KERN_ALERT "pgd = %p\n", mm->pgd);
|
||||
pr_alert("pgd = %p\n", mm->pgd);
|
||||
pgd = pgd_offset(mm, addr);
|
||||
printk(KERN_ALERT "[%08lx] *pgd=%08llx",
|
||||
pr_alert("[%08lx] *pgd=%08llx",
|
||||
addr, (long long)pgd_val(*pgd));
|
||||
|
||||
do {
|
||||
@@ -77,31 +77,31 @@ void show_pte(struct mm_struct *mm, unsigned long addr)
|
||||
break;
|
||||
|
||||
if (pgd_bad(*pgd)) {
|
||||
printk("(bad)");
|
||||
pr_cont("(bad)");
|
||||
break;
|
||||
}
|
||||
|
||||
pud = pud_offset(pgd, addr);
|
||||
if (PTRS_PER_PUD != 1)
|
||||
printk(", *pud=%08llx", (long long)pud_val(*pud));
|
||||
pr_cont(", *pud=%08llx", (long long)pud_val(*pud));
|
||||
|
||||
if (pud_none(*pud))
|
||||
break;
|
||||
|
||||
if (pud_bad(*pud)) {
|
||||
printk("(bad)");
|
||||
pr_cont("(bad)");
|
||||
break;
|
||||
}
|
||||
|
||||
pmd = pmd_offset(pud, addr);
|
||||
if (PTRS_PER_PMD != 1)
|
||||
printk(", *pmd=%08llx", (long long)pmd_val(*pmd));
|
||||
pr_cont(", *pmd=%08llx", (long long)pmd_val(*pmd));
|
||||
|
||||
if (pmd_none(*pmd))
|
||||
break;
|
||||
|
||||
if (pmd_bad(*pmd)) {
|
||||
printk("(bad)");
|
||||
pr_cont("(bad)");
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -110,15 +110,15 @@ void show_pte(struct mm_struct *mm, unsigned long addr)
|
||||
break;
|
||||
|
||||
pte = pte_offset_map(pmd, addr);
|
||||
printk(", *pte=%08llx", (long long)pte_val(*pte));
|
||||
pr_cont(", *pte=%08llx", (long long)pte_val(*pte));
|
||||
#ifndef CONFIG_ARM_LPAE
|
||||
printk(", *ppte=%08llx",
|
||||
pr_cont(", *ppte=%08llx",
|
||||
(long long)pte_val(pte[PTE_HWTABLE_PTRS]));
|
||||
#endif
|
||||
pte_unmap(pte);
|
||||
} while(0);
|
||||
|
||||
printk("\n");
|
||||
pr_cont("\n");
|
||||
}
|
||||
#else /* CONFIG_MMU */
|
||||
void show_pte(struct mm_struct *mm, unsigned long addr)
|
||||
@@ -142,10 +142,9 @@ __do_kernel_fault(struct mm_struct *mm, unsigned long addr, unsigned int fsr,
|
||||
* No handler, we'll have to terminate things with extreme prejudice.
|
||||
*/
|
||||
bust_spinlocks(1);
|
||||
printk(KERN_ALERT
|
||||
"Unable to handle kernel %s at virtual address %08lx\n",
|
||||
(addr < PAGE_SIZE) ? "NULL pointer dereference" :
|
||||
"paging request", addr);
|
||||
pr_alert("Unable to handle kernel %s at virtual address %08lx\n",
|
||||
(addr < PAGE_SIZE) ? "NULL pointer dereference" :
|
||||
"paging request", addr);
|
||||
|
||||
show_pte(mm, addr);
|
||||
die("Oops", regs, fsr);
|
||||
@@ -551,7 +550,7 @@ do_DataAbort(unsigned long addr, unsigned int fsr, struct pt_regs *regs)
|
||||
if (!inf->fn(addr, fsr & ~FSR_LNX_PF, regs))
|
||||
return;
|
||||
|
||||
printk(KERN_ALERT "Unhandled fault: %s (0x%03x) at 0x%08lx\n",
|
||||
pr_alert("Unhandled fault: %s (0x%03x) at 0x%08lx\n",
|
||||
inf->name, fsr, addr);
|
||||
|
||||
info.si_signo = inf->sig;
|
||||
@@ -583,7 +582,7 @@ do_PrefetchAbort(unsigned long addr, unsigned int ifsr, struct pt_regs *regs)
|
||||
if (!inf->fn(addr, ifsr | FSR_LNX_PF, regs))
|
||||
return;
|
||||
|
||||
printk(KERN_ALERT "Unhandled prefetch abort: %s (0x%03x) at 0x%08lx\n",
|
||||
pr_alert("Unhandled prefetch abort: %s (0x%03x) at 0x%08lx\n",
|
||||
inf->name, ifsr, addr);
|
||||
|
||||
info.si_signo = inf->sig;
|
||||
|
@@ -67,7 +67,7 @@ early_param("initrd", early_initrd);
|
||||
|
||||
static int __init parse_tag_initrd(const struct tag *tag)
|
||||
{
|
||||
printk(KERN_WARNING "ATAG_INITRD is deprecated; "
|
||||
pr_warn("ATAG_INITRD is deprecated; "
|
||||
"please update your bootloader.\n");
|
||||
phys_initrd_start = __virt_to_phys(tag->u.initrd.start);
|
||||
phys_initrd_size = tag->u.initrd.size;
|
||||
@@ -544,7 +544,7 @@ void __init mem_init(void)
|
||||
#define MLM(b, t) b, t, ((t) - (b)) >> 20
|
||||
#define MLK_ROUNDUP(b, t) b, t, DIV_ROUND_UP(((t) - (b)), SZ_1K)
|
||||
|
||||
printk(KERN_NOTICE "Virtual kernel memory layout:\n"
|
||||
pr_notice("Virtual kernel memory layout:\n"
|
||||
" vector : 0x%08lx - 0x%08lx (%4ld kB)\n"
|
||||
#ifdef CONFIG_HAVE_TCM
|
||||
" DTCM : 0x%08lx - 0x%08lx (%4ld kB)\n"
|
||||
|
@@ -192,7 +192,7 @@ early_param("cachepolicy", early_cachepolicy);
|
||||
static int __init early_nocache(char *__unused)
|
||||
{
|
||||
char *p = "buffered";
|
||||
printk(KERN_WARNING "nocache is deprecated; use cachepolicy=%s\n", p);
|
||||
pr_warn("nocache is deprecated; use cachepolicy=%s\n", p);
|
||||
early_cachepolicy(p);
|
||||
return 0;
|
||||
}
|
||||
@@ -201,7 +201,7 @@ early_param("nocache", early_nocache);
|
||||
static int __init early_nowrite(char *__unused)
|
||||
{
|
||||
char *p = "uncached";
|
||||
printk(KERN_WARNING "nowb is deprecated; use cachepolicy=%s\n", p);
|
||||
pr_warn("nowb is deprecated; use cachepolicy=%s\n", p);
|
||||
early_cachepolicy(p);
|
||||
return 0;
|
||||
}
|
||||
@@ -786,8 +786,7 @@ static void __init create_36bit_mapping(struct map_desc *md,
|
||||
length = PAGE_ALIGN(md->length);
|
||||
|
||||
if (!(cpu_architecture() >= CPU_ARCH_ARMv6 || cpu_is_xsc3())) {
|
||||
printk(KERN_ERR "MM: CPU does not support supersection "
|
||||
"mapping for 0x%08llx at 0x%08lx\n",
|
||||
pr_err("MM: CPU does not support supersection mapping for 0x%08llx at 0x%08lx\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), addr);
|
||||
return;
|
||||
}
|
||||
@@ -799,15 +798,13 @@ static void __init create_36bit_mapping(struct map_desc *md,
|
||||
* of the actual domain assignments in use.
|
||||
*/
|
||||
if (type->domain) {
|
||||
printk(KERN_ERR "MM: invalid domain in supersection "
|
||||
"mapping for 0x%08llx at 0x%08lx\n",
|
||||
pr_err("MM: invalid domain in supersection mapping for 0x%08llx at 0x%08lx\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), addr);
|
||||
return;
|
||||
}
|
||||
|
||||
if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) {
|
||||
printk(KERN_ERR "MM: cannot create mapping for 0x%08llx"
|
||||
" at 0x%08lx invalid alignment\n",
|
||||
pr_err("MM: cannot create mapping for 0x%08llx at 0x%08lx invalid alignment\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), addr);
|
||||
return;
|
||||
}
|
||||
@@ -850,18 +847,16 @@ static void __init create_mapping(struct map_desc *md)
|
||||
pgd_t *pgd;
|
||||
|
||||
if (md->virtual != vectors_base() && md->virtual < TASK_SIZE) {
|
||||
printk(KERN_WARNING "BUG: not creating mapping for 0x%08llx"
|
||||
" at 0x%08lx in user region\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), md->virtual);
|
||||
pr_warn("BUG: not creating mapping for 0x%08llx at 0x%08lx in user region\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), md->virtual);
|
||||
return;
|
||||
}
|
||||
|
||||
if ((md->type == MT_DEVICE || md->type == MT_ROM) &&
|
||||
md->virtual >= PAGE_OFFSET &&
|
||||
(md->virtual < VMALLOC_START || md->virtual >= VMALLOC_END)) {
|
||||
printk(KERN_WARNING "BUG: mapping for 0x%08llx"
|
||||
" at 0x%08lx out of vmalloc space\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), md->virtual);
|
||||
pr_warn("BUG: mapping for 0x%08llx at 0x%08lx out of vmalloc space\n",
|
||||
(long long)__pfn_to_phys((u64)md->pfn), md->virtual);
|
||||
}
|
||||
|
||||
type = &mem_types[md->type];
|
||||
@@ -881,9 +876,8 @@ static void __init create_mapping(struct map_desc *md)
|
||||
length = PAGE_ALIGN(md->length + (md->virtual & ~PAGE_MASK));
|
||||
|
||||
if (type->prot_l1 == 0 && ((addr | phys | length) & ~SECTION_MASK)) {
|
||||
printk(KERN_WARNING "BUG: map for 0x%08llx at 0x%08lx can not "
|
||||
"be mapped using pages, ignoring.\n",
|
||||
(long long)__pfn_to_phys(md->pfn), addr);
|
||||
pr_warn("BUG: map for 0x%08llx at 0x%08lx can not be mapped using pages, ignoring.\n",
|
||||
(long long)__pfn_to_phys(md->pfn), addr);
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -1053,15 +1047,13 @@ static int __init early_vmalloc(char *arg)
|
||||
|
||||
if (vmalloc_reserve < SZ_16M) {
|
||||
vmalloc_reserve = SZ_16M;
|
||||
printk(KERN_WARNING
|
||||
"vmalloc area too small, limiting to %luMB\n",
|
||||
pr_warn("vmalloc area too small, limiting to %luMB\n",
|
||||
vmalloc_reserve >> 20);
|
||||
}
|
||||
|
||||
if (vmalloc_reserve > VMALLOC_END - (PAGE_OFFSET + SZ_32M)) {
|
||||
vmalloc_reserve = VMALLOC_END - (PAGE_OFFSET + SZ_32M);
|
||||
printk(KERN_WARNING
|
||||
"vmalloc area is too big, limiting to %luMB\n",
|
||||
pr_warn("vmalloc area is too big, limiting to %luMB\n",
|
||||
vmalloc_reserve >> 20);
|
||||
}
|
||||
|
||||
@@ -1094,7 +1086,7 @@ void __init sanity_check_meminfo(void)
|
||||
|
||||
if (highmem) {
|
||||
pr_notice("Ignoring RAM at %pa-%pa (!CONFIG_HIGHMEM)\n",
|
||||
&block_start, &block_end);
|
||||
&block_start, &block_end);
|
||||
memblock_remove(reg->base, reg->size);
|
||||
continue;
|
||||
}
|
||||
@@ -1103,7 +1095,7 @@ void __init sanity_check_meminfo(void)
|
||||
phys_addr_t overlap_size = reg->size - size_limit;
|
||||
|
||||
pr_notice("Truncating RAM at %pa-%pa to -%pa",
|
||||
&block_start, &block_end, &vmalloc_limit);
|
||||
&block_start, &block_end, &vmalloc_limit);
|
||||
memblock_remove(vmalloc_limit, overlap_size);
|
||||
block_end = vmalloc_limit;
|
||||
}
|
||||
|
Criar uma nova questão referindo esta
Bloquear um utilizador