Merge branch 'for-linus' of git://git.armlinux.org.uk/~rmk/linux-arm
Pull ARM updates from Russell King: - StrongARM SA1111 updates to modernise and remove cruft - Add StrongARM gpio drivers for board GPIOs - Verify size of zImage is what we expect to avoid issues with appended DTB - nommu updates from Vladimir Murzin - page table read-write-execute checking from Jinbum Park - Broadcom Brahma-B15 cache updates from Florian Fainelli - Avoid failure with kprobes test caused by inappropriately placed kprobes - Remove __memzero optimisation (which was incorrectly being used directly by some drivers) * 'for-linus' of git://git.armlinux.org.uk/~rmk/linux-arm: (32 commits) ARM: 8745/1: get rid of __memzero() ARM: 8744/1: don't discard memblock for kexec ARM: 8743/1: bL_switcher: add MODULE_LICENSE tag ARM: 8742/1: Always use REFCOUNT_FULL ARM: 8741/1: B15: fix unused label warnings ARM: 8740/1: NOMMU: Make sure we do not hold stale data in mem[] array ARM: 8739/1: NOMMU: Setup VBAR/Hivecs for secondaries cores ARM: 8738/1: Disable CONFIG_DEBUG_VIRTUAL for NOMMU ARM: 8737/1: mm: dump: add checking for writable and executable ARM: 8736/1: mm: dump: make the page table dumping seq_file ARM: 8735/1: mm: dump: make page table dumping reusable ARM: sa1100/neponset: add GPIO drivers for control and modem registers ARM: sa1100/assabet: add BCR/BSR GPIO driver ARM: 8734/1: mm: idmap: Mark variables as ro_after_init ARM: 8733/1: hw_breakpoint: Mark variables as __ro_after_init ARM: 8732/1: NOMMU: Allow userspace to access background MPU region ARM: 8727/1: MAINTAINERS: Update brcmstb entries to cover B15 code ARM: 8728/1: B15: Register reboot notifier for KEXEC ARM: 8730/1: B15: Add suspend/resume hooks ARM: 8726/1: B15: Add CPU hotplug awareness ...
This commit is contained in:
@@ -909,6 +909,14 @@ config OUTER_CACHE_SYNC
|
||||
The outer cache has a outer_cache_fns.sync function pointer
|
||||
that can be used to drain the write buffer of the outer cache.
|
||||
|
||||
config CACHE_B15_RAC
|
||||
bool "Enable the Broadcom Brahma-B15 read-ahead cache controller"
|
||||
depends on ARCH_BRCMSTB
|
||||
default y
|
||||
help
|
||||
This option enables the Broadcom Brahma-B15 read-ahead cache
|
||||
controller. If disabled, the read-ahead cache remains off.
|
||||
|
||||
config CACHE_FEROCEON_L2
|
||||
bool "Enable the Feroceon L2 cache controller"
|
||||
depends on ARCH_MV78XX0 || ARCH_MVEBU
|
||||
|
@@ -13,7 +13,8 @@ obj-y += nommu.o
|
||||
obj-$(CONFIG_ARM_MPU) += pmsa-v7.o
|
||||
endif
|
||||
|
||||
obj-$(CONFIG_ARM_PTDUMP) += dump.o
|
||||
obj-$(CONFIG_ARM_PTDUMP_CORE) += dump.o
|
||||
obj-$(CONFIG_ARM_PTDUMP_DEBUGFS) += ptdump_debugfs.o
|
||||
obj-$(CONFIG_MODULES) += proc-syms.o
|
||||
obj-$(CONFIG_DEBUG_VIRTUAL) += physaddr.o
|
||||
|
||||
@@ -103,6 +104,7 @@ AFLAGS_proc-v6.o :=-Wa,-march=armv6
|
||||
AFLAGS_proc-v7.o :=-Wa,-march=armv7-a
|
||||
|
||||
obj-$(CONFIG_OUTER_CACHE) += l2c-common.o
|
||||
obj-$(CONFIG_CACHE_B15_RAC) += cache-b15-rac.o
|
||||
obj-$(CONFIG_CACHE_FEROCEON_L2) += cache-feroceon-l2.o
|
||||
obj-$(CONFIG_CACHE_L2X0) += cache-l2x0.o l2c-l2x0-resume.o
|
||||
obj-$(CONFIG_CACHE_L2X0_PMU) += cache-l2x0-pmu.o
|
||||
|
356
arch/arm/mm/cache-b15-rac.c
Normal file
356
arch/arm/mm/cache-b15-rac.c
Normal file
@@ -0,0 +1,356 @@
|
||||
/*
|
||||
* Broadcom Brahma-B15 CPU read-ahead cache management functions
|
||||
*
|
||||
* Copyright (C) 2015-2016 Broadcom
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify
|
||||
* it under the terms of the GNU General Public License version 2 as
|
||||
* published by the Free Software Foundation.
|
||||
*/
|
||||
|
||||
#include <linux/err.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/io.h>
|
||||
#include <linux/bitops.h>
|
||||
#include <linux/of_address.h>
|
||||
#include <linux/notifier.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/syscore_ops.h>
|
||||
#include <linux/reboot.h>
|
||||
|
||||
#include <asm/cacheflush.h>
|
||||
#include <asm/hardware/cache-b15-rac.h>
|
||||
|
||||
extern void v7_flush_kern_cache_all(void);
|
||||
|
||||
/* RAC register offsets, relative to the HIF_CPU_BIUCTRL register base */
|
||||
#define RAC_CONFIG0_REG (0x78)
|
||||
#define RACENPREF_MASK (0x3)
|
||||
#define RACPREFINST_SHIFT (0)
|
||||
#define RACENINST_SHIFT (2)
|
||||
#define RACPREFDATA_SHIFT (4)
|
||||
#define RACENDATA_SHIFT (6)
|
||||
#define RAC_CPU_SHIFT (8)
|
||||
#define RACCFG_MASK (0xff)
|
||||
#define RAC_CONFIG1_REG (0x7c)
|
||||
#define RAC_FLUSH_REG (0x80)
|
||||
#define FLUSH_RAC (1 << 0)
|
||||
|
||||
/* Bitmask to enable instruction and data prefetching with a 256-bytes stride */
|
||||
#define RAC_DATA_INST_EN_MASK (1 << RACPREFINST_SHIFT | \
|
||||
RACENPREF_MASK << RACENINST_SHIFT | \
|
||||
1 << RACPREFDATA_SHIFT | \
|
||||
RACENPREF_MASK << RACENDATA_SHIFT)
|
||||
|
||||
#define RAC_ENABLED 0
|
||||
/* Special state where we want to bypass the spinlock and call directly
|
||||
* into the v7 cache maintenance operations during suspend/resume
|
||||
*/
|
||||
#define RAC_SUSPENDED 1
|
||||
|
||||
static void __iomem *b15_rac_base;
|
||||
static DEFINE_SPINLOCK(rac_lock);
|
||||
|
||||
static u32 rac_config0_reg;
|
||||
|
||||
/* Initialization flag to avoid checking for b15_rac_base, and to prevent
|
||||
* multi-platform kernels from crashing here as well.
|
||||
*/
|
||||
static unsigned long b15_rac_flags;
|
||||
|
||||
static inline u32 __b15_rac_disable(void)
|
||||
{
|
||||
u32 val = __raw_readl(b15_rac_base + RAC_CONFIG0_REG);
|
||||
__raw_writel(0, b15_rac_base + RAC_CONFIG0_REG);
|
||||
dmb();
|
||||
return val;
|
||||
}
|
||||
|
||||
static inline void __b15_rac_flush(void)
|
||||
{
|
||||
u32 reg;
|
||||
|
||||
__raw_writel(FLUSH_RAC, b15_rac_base + RAC_FLUSH_REG);
|
||||
do {
|
||||
/* This dmb() is required to force the Bus Interface Unit
|
||||
* to clean oustanding writes, and forces an idle cycle
|
||||
* to be inserted.
|
||||
*/
|
||||
dmb();
|
||||
reg = __raw_readl(b15_rac_base + RAC_FLUSH_REG);
|
||||
} while (reg & FLUSH_RAC);
|
||||
}
|
||||
|
||||
static inline u32 b15_rac_disable_and_flush(void)
|
||||
{
|
||||
u32 reg;
|
||||
|
||||
reg = __b15_rac_disable();
|
||||
__b15_rac_flush();
|
||||
return reg;
|
||||
}
|
||||
|
||||
static inline void __b15_rac_enable(u32 val)
|
||||
{
|
||||
__raw_writel(val, b15_rac_base + RAC_CONFIG0_REG);
|
||||
/* dsb() is required here to be consistent with __flush_icache_all() */
|
||||
dsb();
|
||||
}
|
||||
|
||||
#define BUILD_RAC_CACHE_OP(name, bar) \
|
||||
void b15_flush_##name(void) \
|
||||
{ \
|
||||
unsigned int do_flush; \
|
||||
u32 val = 0; \
|
||||
\
|
||||
if (test_bit(RAC_SUSPENDED, &b15_rac_flags)) { \
|
||||
v7_flush_##name(); \
|
||||
bar; \
|
||||
return; \
|
||||
} \
|
||||
\
|
||||
spin_lock(&rac_lock); \
|
||||
do_flush = test_bit(RAC_ENABLED, &b15_rac_flags); \
|
||||
if (do_flush) \
|
||||
val = b15_rac_disable_and_flush(); \
|
||||
v7_flush_##name(); \
|
||||
if (!do_flush) \
|
||||
bar; \
|
||||
else \
|
||||
__b15_rac_enable(val); \
|
||||
spin_unlock(&rac_lock); \
|
||||
}
|
||||
|
||||
#define nobarrier
|
||||
|
||||
/* The readahead cache present in the Brahma-B15 CPU is a special piece of
|
||||
* hardware after the integrated L2 cache of the B15 CPU complex whose purpose
|
||||
* is to prefetch instruction and/or data with a line size of either 64 bytes
|
||||
* or 256 bytes. The rationale is that the data-bus of the CPU interface is
|
||||
* optimized for 256-bytes transactions, and enabling the readahead cache
|
||||
* provides a significant performance boost we want it enabled (typically
|
||||
* twice the performance for a memcpy benchmark application).
|
||||
*
|
||||
* The readahead cache is transparent for Modified Virtual Addresses
|
||||
* cache maintenance operations: ICIMVAU, DCIMVAC, DCCMVAC, DCCMVAU and
|
||||
* DCCIMVAC.
|
||||
*
|
||||
* It is however not transparent for the following cache maintenance
|
||||
* operations: DCISW, DCCSW, DCCISW, ICIALLUIS and ICIALLU which is precisely
|
||||
* what we are patching here with our BUILD_RAC_CACHE_OP here.
|
||||
*/
|
||||
BUILD_RAC_CACHE_OP(kern_cache_all, nobarrier);
|
||||
|
||||
static void b15_rac_enable(void)
|
||||
{
|
||||
unsigned int cpu;
|
||||
u32 enable = 0;
|
||||
|
||||
for_each_possible_cpu(cpu)
|
||||
enable |= (RAC_DATA_INST_EN_MASK << (cpu * RAC_CPU_SHIFT));
|
||||
|
||||
b15_rac_disable_and_flush();
|
||||
__b15_rac_enable(enable);
|
||||
}
|
||||
|
||||
static int b15_rac_reboot_notifier(struct notifier_block *nb,
|
||||
unsigned long action,
|
||||
void *data)
|
||||
{
|
||||
/* During kexec, we are not yet migrated on the boot CPU, so we need to
|
||||
* make sure we are SMP safe here. Once the RAC is disabled, flag it as
|
||||
* suspended such that the hotplug notifier returns early.
|
||||
*/
|
||||
if (action == SYS_RESTART) {
|
||||
spin_lock(&rac_lock);
|
||||
b15_rac_disable_and_flush();
|
||||
clear_bit(RAC_ENABLED, &b15_rac_flags);
|
||||
set_bit(RAC_SUSPENDED, &b15_rac_flags);
|
||||
spin_unlock(&rac_lock);
|
||||
}
|
||||
|
||||
return NOTIFY_DONE;
|
||||
}
|
||||
|
||||
static struct notifier_block b15_rac_reboot_nb = {
|
||||
.notifier_call = b15_rac_reboot_notifier,
|
||||
};
|
||||
|
||||
/* The CPU hotplug case is the most interesting one, we basically need to make
|
||||
* sure that the RAC is disabled for the entire system prior to having a CPU
|
||||
* die, in particular prior to this dying CPU having exited the coherency
|
||||
* domain.
|
||||
*
|
||||
* Once this CPU is marked dead, we can safely re-enable the RAC for the
|
||||
* remaining CPUs in the system which are still online.
|
||||
*
|
||||
* Offlining a CPU is the problematic case, onlining a CPU is not much of an
|
||||
* issue since the CPU and its cache-level hierarchy will start filling with
|
||||
* the RAC disabled, so L1 and L2 only.
|
||||
*
|
||||
* In this function, we should NOT have to verify any unsafe setting/condition
|
||||
* b15_rac_base:
|
||||
*
|
||||
* It is protected by the RAC_ENABLED flag which is cleared by default, and
|
||||
* being cleared when initial procedure is done. b15_rac_base had been set at
|
||||
* that time.
|
||||
*
|
||||
* RAC_ENABLED:
|
||||
* There is a small timing windows, in b15_rac_init(), between
|
||||
* cpuhp_setup_state_*()
|
||||
* ...
|
||||
* set RAC_ENABLED
|
||||
* However, there is no hotplug activity based on the Linux booting procedure.
|
||||
*
|
||||
* Since we have to disable RAC for all cores, we keep RAC on as long as as
|
||||
* possible (disable it as late as possible) to gain the cache benefit.
|
||||
*
|
||||
* Thus, dying/dead states are chosen here
|
||||
*
|
||||
* We are choosing not do disable the RAC on a per-CPU basis, here, if we did
|
||||
* we would want to consider disabling it as early as possible to benefit the
|
||||
* other active CPUs.
|
||||
*/
|
||||
|
||||
/* Running on the dying CPU */
|
||||
static int b15_rac_dying_cpu(unsigned int cpu)
|
||||
{
|
||||
/* During kexec/reboot, the RAC is disabled via the reboot notifier
|
||||
* return early here.
|
||||
*/
|
||||
if (test_bit(RAC_SUSPENDED, &b15_rac_flags))
|
||||
return 0;
|
||||
|
||||
spin_lock(&rac_lock);
|
||||
|
||||
/* Indicate that we are starting a hotplug procedure */
|
||||
__clear_bit(RAC_ENABLED, &b15_rac_flags);
|
||||
|
||||
/* Disable the readahead cache and save its value to a global */
|
||||
rac_config0_reg = b15_rac_disable_and_flush();
|
||||
|
||||
spin_unlock(&rac_lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* Running on a non-dying CPU */
|
||||
static int b15_rac_dead_cpu(unsigned int cpu)
|
||||
{
|
||||
/* During kexec/reboot, the RAC is disabled via the reboot notifier
|
||||
* return early here.
|
||||
*/
|
||||
if (test_bit(RAC_SUSPENDED, &b15_rac_flags))
|
||||
return 0;
|
||||
|
||||
spin_lock(&rac_lock);
|
||||
|
||||
/* And enable it */
|
||||
__b15_rac_enable(rac_config0_reg);
|
||||
__set_bit(RAC_ENABLED, &b15_rac_flags);
|
||||
|
||||
spin_unlock(&rac_lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int b15_rac_suspend(void)
|
||||
{
|
||||
/* Suspend the read-ahead cache oeprations, forcing our cache
|
||||
* implementation to fallback to the regular ARMv7 calls.
|
||||
*
|
||||
* We are guaranteed to be running on the boot CPU at this point and
|
||||
* with every other CPU quiesced, so setting RAC_SUSPENDED is not racy
|
||||
* here.
|
||||
*/
|
||||
rac_config0_reg = b15_rac_disable_and_flush();
|
||||
set_bit(RAC_SUSPENDED, &b15_rac_flags);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void b15_rac_resume(void)
|
||||
{
|
||||
/* Coming out of a S3 suspend/resume cycle, the read-ahead cache
|
||||
* register RAC_CONFIG0_REG will be restored to its default value, make
|
||||
* sure we re-enable it and set the enable flag, we are also guaranteed
|
||||
* to run on the boot CPU, so not racy again.
|
||||
*/
|
||||
__b15_rac_enable(rac_config0_reg);
|
||||
clear_bit(RAC_SUSPENDED, &b15_rac_flags);
|
||||
}
|
||||
|
||||
static struct syscore_ops b15_rac_syscore_ops = {
|
||||
.suspend = b15_rac_suspend,
|
||||
.resume = b15_rac_resume,
|
||||
};
|
||||
|
||||
static int __init b15_rac_init(void)
|
||||
{
|
||||
struct device_node *dn;
|
||||
int ret = 0, cpu;
|
||||
u32 reg, en_mask = 0;
|
||||
|
||||
dn = of_find_compatible_node(NULL, NULL, "brcm,brcmstb-cpu-biu-ctrl");
|
||||
if (!dn)
|
||||
return -ENODEV;
|
||||
|
||||
if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
|
||||
goto out;
|
||||
|
||||
b15_rac_base = of_iomap(dn, 0);
|
||||
if (!b15_rac_base) {
|
||||
pr_err("failed to remap BIU control base\n");
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
ret = register_reboot_notifier(&b15_rac_reboot_nb);
|
||||
if (ret) {
|
||||
pr_err("failed to register reboot notifier\n");
|
||||
iounmap(b15_rac_base);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (IS_ENABLED(CONFIG_HOTPLUG_CPU)) {
|
||||
ret = cpuhp_setup_state_nocalls(CPUHP_AP_ARM_CACHE_B15_RAC_DEAD,
|
||||
"arm/cache-b15-rac:dead",
|
||||
NULL, b15_rac_dead_cpu);
|
||||
if (ret)
|
||||
goto out_unmap;
|
||||
|
||||
ret = cpuhp_setup_state_nocalls(CPUHP_AP_ARM_CACHE_B15_RAC_DYING,
|
||||
"arm/cache-b15-rac:dying",
|
||||
NULL, b15_rac_dying_cpu);
|
||||
if (ret)
|
||||
goto out_cpu_dead;
|
||||
}
|
||||
|
||||
if (IS_ENABLED(CONFIG_PM_SLEEP))
|
||||
register_syscore_ops(&b15_rac_syscore_ops);
|
||||
|
||||
spin_lock(&rac_lock);
|
||||
reg = __raw_readl(b15_rac_base + RAC_CONFIG0_REG);
|
||||
for_each_possible_cpu(cpu)
|
||||
en_mask |= ((1 << RACPREFDATA_SHIFT) << (cpu * RAC_CPU_SHIFT));
|
||||
WARN(reg & en_mask, "Read-ahead cache not previously disabled\n");
|
||||
|
||||
b15_rac_enable();
|
||||
set_bit(RAC_ENABLED, &b15_rac_flags);
|
||||
spin_unlock(&rac_lock);
|
||||
|
||||
pr_info("Broadcom Brahma-B15 readahead cache at: 0x%p\n",
|
||||
b15_rac_base + RAC_CONFIG0_REG);
|
||||
|
||||
goto out;
|
||||
|
||||
out_cpu_dead:
|
||||
cpuhp_remove_state_nocalls(CPUHP_AP_ARM_CACHE_B15_RAC_DYING);
|
||||
out_unmap:
|
||||
unregister_reboot_notifier(&b15_rac_reboot_nb);
|
||||
iounmap(b15_rac_base);
|
||||
out:
|
||||
of_node_put(dn);
|
||||
return ret;
|
||||
}
|
||||
arch_initcall(b15_rac_init);
|
@@ -15,6 +15,7 @@
|
||||
#include <asm/assembler.h>
|
||||
#include <asm/errno.h>
|
||||
#include <asm/unwind.h>
|
||||
#include <asm/hardware/cache-b15-rac.h>
|
||||
|
||||
#include "proc-macros.S"
|
||||
|
||||
@@ -446,3 +447,23 @@ ENDPROC(v7_dma_unmap_area)
|
||||
|
||||
@ define struct cpu_cache_fns (see <asm/cacheflush.h> and proc-macros.S)
|
||||
define_cache_functions v7
|
||||
|
||||
/* The Broadcom Brahma-B15 read-ahead cache requires some modifications
|
||||
* to the v7_cache_fns, we only override the ones we need
|
||||
*/
|
||||
#ifndef CONFIG_CACHE_B15_RAC
|
||||
globl_equ b15_flush_kern_cache_all, v7_flush_kern_cache_all
|
||||
#endif
|
||||
globl_equ b15_flush_icache_all, v7_flush_icache_all
|
||||
globl_equ b15_flush_kern_cache_louis, v7_flush_kern_cache_louis
|
||||
globl_equ b15_flush_user_cache_all, v7_flush_user_cache_all
|
||||
globl_equ b15_flush_user_cache_range, v7_flush_user_cache_range
|
||||
globl_equ b15_coherent_kern_range, v7_coherent_kern_range
|
||||
globl_equ b15_coherent_user_range, v7_coherent_user_range
|
||||
globl_equ b15_flush_kern_dcache_area, v7_flush_kern_dcache_area
|
||||
|
||||
globl_equ b15_dma_map_area, v7_dma_map_area
|
||||
globl_equ b15_dma_unmap_area, v7_dma_unmap_area
|
||||
globl_equ b15_dma_flush_range, v7_dma_flush_range
|
||||
|
||||
define_cache_functions b15
|
||||
|
@@ -21,11 +21,7 @@
|
||||
#include <asm/fixmap.h>
|
||||
#include <asm/memory.h>
|
||||
#include <asm/pgtable.h>
|
||||
|
||||
struct addr_marker {
|
||||
unsigned long start_address;
|
||||
const char *name;
|
||||
};
|
||||
#include <asm/ptdump.h>
|
||||
|
||||
static struct addr_marker address_markers[] = {
|
||||
{ MODULES_VADDR, "Modules" },
|
||||
@@ -38,12 +34,26 @@ static struct addr_marker address_markers[] = {
|
||||
{ -1, NULL },
|
||||
};
|
||||
|
||||
#define pt_dump_seq_printf(m, fmt, args...) \
|
||||
({ \
|
||||
if (m) \
|
||||
seq_printf(m, fmt, ##args); \
|
||||
})
|
||||
|
||||
#define pt_dump_seq_puts(m, fmt) \
|
||||
({ \
|
||||
if (m) \
|
||||
seq_printf(m, fmt); \
|
||||
})
|
||||
|
||||
struct pg_state {
|
||||
struct seq_file *seq;
|
||||
const struct addr_marker *marker;
|
||||
unsigned long start_address;
|
||||
unsigned level;
|
||||
u64 current_prot;
|
||||
bool check_wx;
|
||||
unsigned long wx_pages;
|
||||
const char *current_domain;
|
||||
};
|
||||
|
||||
@@ -52,6 +62,8 @@ struct prot_bits {
|
||||
u64 val;
|
||||
const char *set;
|
||||
const char *clear;
|
||||
bool ro_bit;
|
||||
bool nx_bit;
|
||||
};
|
||||
|
||||
static const struct prot_bits pte_bits[] = {
|
||||
@@ -65,11 +77,13 @@ static const struct prot_bits pte_bits[] = {
|
||||
.val = L_PTE_RDONLY,
|
||||
.set = "ro",
|
||||
.clear = "RW",
|
||||
.ro_bit = true,
|
||||
}, {
|
||||
.mask = L_PTE_XN,
|
||||
.val = L_PTE_XN,
|
||||
.set = "NX",
|
||||
.clear = "x ",
|
||||
.nx_bit = true,
|
||||
}, {
|
||||
.mask = L_PTE_SHARED,
|
||||
.val = L_PTE_SHARED,
|
||||
@@ -133,11 +147,13 @@ static const struct prot_bits section_bits[] = {
|
||||
.val = L_PMD_SECT_RDONLY | PMD_SECT_AP2,
|
||||
.set = "ro",
|
||||
.clear = "RW",
|
||||
.ro_bit = true,
|
||||
#elif __LINUX_ARM_ARCH__ >= 6
|
||||
{
|
||||
.mask = PMD_SECT_APX | PMD_SECT_AP_READ | PMD_SECT_AP_WRITE,
|
||||
.val = PMD_SECT_APX | PMD_SECT_AP_WRITE,
|
||||
.set = " ro",
|
||||
.ro_bit = true,
|
||||
}, {
|
||||
.mask = PMD_SECT_APX | PMD_SECT_AP_READ | PMD_SECT_AP_WRITE,
|
||||
.val = PMD_SECT_AP_WRITE,
|
||||
@@ -156,6 +172,7 @@ static const struct prot_bits section_bits[] = {
|
||||
.mask = PMD_SECT_AP_READ | PMD_SECT_AP_WRITE,
|
||||
.val = 0,
|
||||
.set = " ro",
|
||||
.ro_bit = true,
|
||||
}, {
|
||||
.mask = PMD_SECT_AP_READ | PMD_SECT_AP_WRITE,
|
||||
.val = PMD_SECT_AP_WRITE,
|
||||
@@ -174,6 +191,7 @@ static const struct prot_bits section_bits[] = {
|
||||
.val = PMD_SECT_XN,
|
||||
.set = "NX",
|
||||
.clear = "x ",
|
||||
.nx_bit = true,
|
||||
}, {
|
||||
.mask = PMD_SECT_S,
|
||||
.val = PMD_SECT_S,
|
||||
@@ -186,6 +204,8 @@ struct pg_level {
|
||||
const struct prot_bits *bits;
|
||||
size_t num;
|
||||
u64 mask;
|
||||
const struct prot_bits *ro_bit;
|
||||
const struct prot_bits *nx_bit;
|
||||
};
|
||||
|
||||
static struct pg_level pg_level[] = {
|
||||
@@ -214,10 +234,27 @@ static void dump_prot(struct pg_state *st, const struct prot_bits *bits, size_t
|
||||
s = bits->clear;
|
||||
|
||||
if (s)
|
||||
seq_printf(st->seq, " %s", s);
|
||||
pt_dump_seq_printf(st->seq, " %s", s);
|
||||
}
|
||||
}
|
||||
|
||||
static void note_prot_wx(struct pg_state *st, unsigned long addr)
|
||||
{
|
||||
if (!st->check_wx)
|
||||
return;
|
||||
if ((st->current_prot & pg_level[st->level].ro_bit->mask) ==
|
||||
pg_level[st->level].ro_bit->val)
|
||||
return;
|
||||
if ((st->current_prot & pg_level[st->level].nx_bit->mask) ==
|
||||
pg_level[st->level].nx_bit->val)
|
||||
return;
|
||||
|
||||
WARN_ONCE(1, "arm/mm: Found insecure W+X mapping at address %pS\n",
|
||||
(void *)st->start_address);
|
||||
|
||||
st->wx_pages += (addr - st->start_address) / PAGE_SIZE;
|
||||
}
|
||||
|
||||
static void note_page(struct pg_state *st, unsigned long addr,
|
||||
unsigned int level, u64 val, const char *domain)
|
||||
{
|
||||
@@ -228,7 +265,7 @@ static void note_page(struct pg_state *st, unsigned long addr,
|
||||
st->level = level;
|
||||
st->current_prot = prot;
|
||||
st->current_domain = domain;
|
||||
seq_printf(st->seq, "---[ %s ]---\n", st->marker->name);
|
||||
pt_dump_seq_printf(st->seq, "---[ %s ]---\n", st->marker->name);
|
||||
} else if (prot != st->current_prot || level != st->level ||
|
||||
domain != st->current_domain ||
|
||||
addr >= st->marker[1].start_address) {
|
||||
@@ -236,7 +273,8 @@ static void note_page(struct pg_state *st, unsigned long addr,
|
||||
unsigned long delta;
|
||||
|
||||
if (st->current_prot) {
|
||||
seq_printf(st->seq, "0x%08lx-0x%08lx ",
|
||||
note_prot_wx(st, addr);
|
||||
pt_dump_seq_printf(st->seq, "0x%08lx-0x%08lx ",
|
||||
st->start_address, addr);
|
||||
|
||||
delta = (addr - st->start_address) >> 10;
|
||||
@@ -244,17 +282,19 @@ static void note_page(struct pg_state *st, unsigned long addr,
|
||||
delta >>= 10;
|
||||
unit++;
|
||||
}
|
||||
seq_printf(st->seq, "%9lu%c", delta, *unit);
|
||||
pt_dump_seq_printf(st->seq, "%9lu%c", delta, *unit);
|
||||
if (st->current_domain)
|
||||
seq_printf(st->seq, " %s", st->current_domain);
|
||||
pt_dump_seq_printf(st->seq, " %s",
|
||||
st->current_domain);
|
||||
if (pg_level[st->level].bits)
|
||||
dump_prot(st, pg_level[st->level].bits, pg_level[st->level].num);
|
||||
seq_printf(st->seq, "\n");
|
||||
pt_dump_seq_printf(st->seq, "\n");
|
||||
}
|
||||
|
||||
if (addr >= st->marker[1].start_address) {
|
||||
st->marker++;
|
||||
seq_printf(st->seq, "---[ %s ]---\n", st->marker->name);
|
||||
pt_dump_seq_printf(st->seq, "---[ %s ]---\n",
|
||||
st->marker->name);
|
||||
}
|
||||
st->start_address = addr;
|
||||
st->current_prot = prot;
|
||||
@@ -335,61 +375,82 @@ static void walk_pud(struct pg_state *st, pgd_t *pgd, unsigned long start)
|
||||
}
|
||||
}
|
||||
|
||||
static void walk_pgd(struct seq_file *m)
|
||||
static void walk_pgd(struct pg_state *st, struct mm_struct *mm,
|
||||
unsigned long start)
|
||||
{
|
||||
pgd_t *pgd = swapper_pg_dir;
|
||||
struct pg_state st;
|
||||
unsigned long addr;
|
||||
pgd_t *pgd = pgd_offset(mm, 0UL);
|
||||
unsigned i;
|
||||
|
||||
memset(&st, 0, sizeof(st));
|
||||
st.seq = m;
|
||||
st.marker = address_markers;
|
||||
unsigned long addr;
|
||||
|
||||
for (i = 0; i < PTRS_PER_PGD; i++, pgd++) {
|
||||
addr = i * PGDIR_SIZE;
|
||||
addr = start + i * PGDIR_SIZE;
|
||||
if (!pgd_none(*pgd)) {
|
||||
walk_pud(&st, pgd, addr);
|
||||
walk_pud(st, pgd, addr);
|
||||
} else {
|
||||
note_page(&st, addr, 1, pgd_val(*pgd), NULL);
|
||||
note_page(st, addr, 1, pgd_val(*pgd), NULL);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void ptdump_walk_pgd(struct seq_file *m, struct ptdump_info *info)
|
||||
{
|
||||
struct pg_state st = {
|
||||
.seq = m,
|
||||
.marker = info->markers,
|
||||
.check_wx = false,
|
||||
};
|
||||
|
||||
walk_pgd(&st, info->mm, info->base_addr);
|
||||
note_page(&st, 0, 0, 0, NULL);
|
||||
}
|
||||
|
||||
static int ptdump_show(struct seq_file *m, void *v)
|
||||
static void ptdump_initialize(void)
|
||||
{
|
||||
walk_pgd(m);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int ptdump_open(struct inode *inode, struct file *file)
|
||||
{
|
||||
return single_open(file, ptdump_show, NULL);
|
||||
}
|
||||
|
||||
static const struct file_operations ptdump_fops = {
|
||||
.open = ptdump_open,
|
||||
.read = seq_read,
|
||||
.llseek = seq_lseek,
|
||||
.release = single_release,
|
||||
};
|
||||
|
||||
static int ptdump_init(void)
|
||||
{
|
||||
struct dentry *pe;
|
||||
unsigned i, j;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(pg_level); i++)
|
||||
if (pg_level[i].bits)
|
||||
for (j = 0; j < pg_level[i].num; j++)
|
||||
for (j = 0; j < pg_level[i].num; j++) {
|
||||
pg_level[i].mask |= pg_level[i].bits[j].mask;
|
||||
if (pg_level[i].bits[j].ro_bit)
|
||||
pg_level[i].ro_bit = &pg_level[i].bits[j];
|
||||
if (pg_level[i].bits[j].nx_bit)
|
||||
pg_level[i].nx_bit = &pg_level[i].bits[j];
|
||||
}
|
||||
|
||||
address_markers[2].start_address = VMALLOC_START;
|
||||
}
|
||||
|
||||
pe = debugfs_create_file("kernel_page_tables", 0400, NULL, NULL,
|
||||
&ptdump_fops);
|
||||
return pe ? 0 : -ENOMEM;
|
||||
static struct ptdump_info kernel_ptdump_info = {
|
||||
.mm = &init_mm,
|
||||
.markers = address_markers,
|
||||
.base_addr = 0,
|
||||
};
|
||||
|
||||
void ptdump_check_wx(void)
|
||||
{
|
||||
struct pg_state st = {
|
||||
.seq = NULL,
|
||||
.marker = (struct addr_marker[]) {
|
||||
{ 0, NULL},
|
||||
{ -1, NULL},
|
||||
},
|
||||
.check_wx = true,
|
||||
};
|
||||
|
||||
walk_pgd(&st, &init_mm, 0);
|
||||
note_page(&st, 0, 0, 0, NULL);
|
||||
if (st.wx_pages)
|
||||
pr_warn("Checked W+X mappings: FAILED, %lu W+X pages found\n",
|
||||
st.wx_pages);
|
||||
else
|
||||
pr_info("Checked W+X mappings: passed, no W+X pages found\n");
|
||||
}
|
||||
|
||||
static int ptdump_init(void)
|
||||
{
|
||||
ptdump_initialize();
|
||||
return ptdump_debugfs_register(&kernel_ptdump_info,
|
||||
"kernel_page_tables");
|
||||
}
|
||||
__initcall(ptdump_init);
|
||||
|
@@ -21,7 +21,6 @@
|
||||
#include <linux/highmem.h>
|
||||
#include <linux/perf_event.h>
|
||||
|
||||
#include <asm/exception.h>
|
||||
#include <asm/pgtable.h>
|
||||
#include <asm/system_misc.h>
|
||||
#include <asm/system_info.h>
|
||||
@@ -545,7 +544,7 @@ hook_fault_code(int nr, int (*fn)(unsigned long, unsigned int, struct pt_regs *)
|
||||
/*
|
||||
* Dispatch a data abort to the relevant handler.
|
||||
*/
|
||||
asmlinkage void __exception
|
||||
asmlinkage void
|
||||
do_DataAbort(unsigned long addr, unsigned int fsr, struct pt_regs *regs)
|
||||
{
|
||||
const struct fsr_info *inf = fsr_info + fsr_fs(fsr);
|
||||
@@ -578,7 +577,7 @@ hook_ifault_code(int nr, int (*fn)(unsigned long, unsigned int, struct pt_regs *
|
||||
ifsr_info[nr].name = name;
|
||||
}
|
||||
|
||||
asmlinkage void __exception
|
||||
asmlinkage void
|
||||
do_PrefetchAbort(unsigned long addr, unsigned int ifsr, struct pt_regs *regs)
|
||||
{
|
||||
const struct fsr_info *inf = ifsr_info + fsr_fs(ifsr);
|
||||
|
@@ -16,8 +16,8 @@
|
||||
* are not supported on any CPU using the idmap tables as its current
|
||||
* page tables.
|
||||
*/
|
||||
pgd_t *idmap_pgd;
|
||||
long long arch_phys_to_idmap_offset;
|
||||
pgd_t *idmap_pgd __ro_after_init;
|
||||
long long arch_phys_to_idmap_offset __ro_after_init;
|
||||
|
||||
#ifdef CONFIG_ARM_LPAE
|
||||
static void idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end,
|
||||
|
@@ -36,6 +36,7 @@
|
||||
#include <asm/system_info.h>
|
||||
#include <asm/tlb.h>
|
||||
#include <asm/fixmap.h>
|
||||
#include <asm/ptdump.h>
|
||||
|
||||
#include <asm/mach/arch.h>
|
||||
#include <asm/mach/map.h>
|
||||
@@ -738,6 +739,7 @@ static int __mark_rodata_ro(void *unused)
|
||||
void mark_rodata_ro(void)
|
||||
{
|
||||
stop_machine(__mark_rodata_ro, NULL, NULL);
|
||||
debug_checkwx();
|
||||
}
|
||||
|
||||
void set_kernel_text_rw(void)
|
||||
|
@@ -31,7 +31,7 @@ struct mpu_rgn_info mpu_rgn_info;
|
||||
|
||||
#ifdef CONFIG_CPU_CP15
|
||||
#ifdef CONFIG_CPU_HIGH_VECTOR
|
||||
static unsigned long __init setup_vectors_base(void)
|
||||
unsigned long setup_vectors_base(void)
|
||||
{
|
||||
unsigned long reg = get_cr();
|
||||
|
||||
@@ -57,7 +57,7 @@ static inline bool security_extensions_enabled(void)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static unsigned long __init setup_vectors_base(void)
|
||||
unsigned long setup_vectors_base(void)
|
||||
{
|
||||
unsigned long base = 0, reg = get_cr();
|
||||
|
||||
|
@@ -6,6 +6,7 @@
|
||||
|
||||
#include <linux/bitops.h>
|
||||
#include <linux/memblock.h>
|
||||
#include <linux/string.h>
|
||||
|
||||
#include <asm/cacheflush.h>
|
||||
#include <asm/cp15.h>
|
||||
@@ -296,6 +297,7 @@ void __init adjust_lowmem_bounds_mpu(void)
|
||||
}
|
||||
}
|
||||
|
||||
memset(mem, 0, sizeof(mem));
|
||||
num = allocate_region(mem_start, specified_mem_size, mem_max_regions, mem);
|
||||
|
||||
for (i = 0; i < num; i++) {
|
||||
@@ -433,7 +435,7 @@ void __init mpu_setup(void)
|
||||
|
||||
/* Background */
|
||||
err |= mpu_setup_region(region++, 0, 32,
|
||||
MPU_ACR_XN | MPU_RGN_STRONGLY_ORDERED | MPU_AP_PL1RW_PL0NA,
|
||||
MPU_ACR_XN | MPU_RGN_STRONGLY_ORDERED | MPU_AP_PL1RW_PL0RW,
|
||||
0, false);
|
||||
|
||||
#ifdef CONFIG_XIP_KERNEL
|
||||
|
@@ -567,7 +567,7 @@ __v7_setup_stack:
|
||||
/*
|
||||
* Standard v7 proc info content
|
||||
*/
|
||||
.macro __v7_proc name, initfunc, mm_mmuflags = 0, io_mmuflags = 0, hwcaps = 0, proc_fns = v7_processor_functions
|
||||
.macro __v7_proc name, initfunc, mm_mmuflags = 0, io_mmuflags = 0, hwcaps = 0, proc_fns = v7_processor_functions, cache_fns = v7_cache_fns
|
||||
ALT_SMP(.long PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_AP_READ | \
|
||||
PMD_SECT_AF | PMD_FLAGS_SMP | \mm_mmuflags)
|
||||
ALT_UP(.long PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_AP_READ | \
|
||||
@@ -583,7 +583,7 @@ __v7_setup_stack:
|
||||
.long \proc_fns
|
||||
.long v7wbi_tlb_fns
|
||||
.long v6_user_fns
|
||||
.long v7_cache_fns
|
||||
.long \cache_fns
|
||||
.endm
|
||||
|
||||
#ifndef CONFIG_ARM_LPAE
|
||||
@@ -678,7 +678,7 @@ __v7_ca15mp_proc_info:
|
||||
__v7_b15mp_proc_info:
|
||||
.long 0x420f00f0
|
||||
.long 0xff0ffff0
|
||||
__v7_proc __v7_b15mp_proc_info, __v7_b15mp_setup
|
||||
__v7_proc __v7_b15mp_proc_info, __v7_b15mp_setup, cache_fns = b15_cache_fns
|
||||
.size __v7_b15mp_proc_info, . - __v7_b15mp_proc_info
|
||||
|
||||
/*
|
||||
|
34
arch/arm/mm/ptdump_debugfs.c
Normal file
34
arch/arm/mm/ptdump_debugfs.c
Normal file
@@ -0,0 +1,34 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
#include <linux/debugfs.h>
|
||||
#include <linux/seq_file.h>
|
||||
|
||||
#include <asm/ptdump.h>
|
||||
|
||||
static int ptdump_show(struct seq_file *m, void *v)
|
||||
{
|
||||
struct ptdump_info *info = m->private;
|
||||
|
||||
ptdump_walk_pgd(m, info);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int ptdump_open(struct inode *inode, struct file *file)
|
||||
{
|
||||
return single_open(file, ptdump_show, inode->i_private);
|
||||
}
|
||||
|
||||
static const struct file_operations ptdump_fops = {
|
||||
.open = ptdump_open,
|
||||
.read = seq_read,
|
||||
.llseek = seq_lseek,
|
||||
.release = single_release,
|
||||
};
|
||||
|
||||
int ptdump_debugfs_register(struct ptdump_info *info, const char *name)
|
||||
{
|
||||
struct dentry *pe;
|
||||
|
||||
pe = debugfs_create_file(name, 0400, NULL, info, &ptdump_fops);
|
||||
return pe ? 0 : -ENOMEM;
|
||||
|
||||
}
|
Reference in New Issue
Block a user