123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186 |
- /* SPDX-License-Identifier: GPL-2.0 */
- /*
- * Copyright (C) 2020 ARM Ltd.
- */
- #ifndef __ASM_MTE_H
- #define __ASM_MTE_H
- #include <asm/compiler.h>
- #include <asm/mte-def.h>
- #ifndef __ASSEMBLY__
- #include <linux/bitfield.h>
- #include <linux/kasan-enabled.h>
- #include <linux/page-flags.h>
- #include <linux/sched.h>
- #include <linux/types.h>
- #include <asm/pgtable-types.h>
- void mte_clear_page_tags(void *addr);
- unsigned long mte_copy_tags_from_user(void *to, const void __user *from,
- unsigned long n);
- unsigned long mte_copy_tags_to_user(void __user *to, void *from,
- unsigned long n);
- int mte_save_tags(struct page *page);
- void mte_save_page_tags(const void *page_addr, void *tag_storage);
- void mte_restore_tags(swp_entry_t entry, struct page *page);
- void mte_restore_page_tags(void *page_addr, const void *tag_storage);
- void mte_invalidate_tags(int type, pgoff_t offset);
- void mte_invalidate_tags_area(int type);
- void *mte_allocate_tag_storage(void);
- void mte_free_tag_storage(char *storage);
- #ifdef CONFIG_ARM64_MTE
- /* track which pages have valid allocation tags */
- #define PG_mte_tagged PG_arch_2
- static inline void set_page_mte_tagged(struct page *page)
- {
- /*
- * Ensure that the tags written prior to this function are visible
- * before the page flags update.
- */
- smp_wmb();
- set_bit(PG_mte_tagged, &page->flags);
- }
- static inline bool page_mte_tagged(struct page *page)
- {
- bool ret = test_bit(PG_mte_tagged, &page->flags);
- /*
- * If the page is tagged, ensure ordering with a likely subsequent
- * read of the tags.
- */
- if (ret)
- smp_rmb();
- return ret;
- }
- void mte_zero_clear_page_tags(void *addr);
- void mte_sync_tags(pte_t pte);
- void mte_copy_page_tags(void *kto, const void *kfrom);
- void mte_thread_init_user(void);
- void mte_thread_switch(struct task_struct *next);
- void mte_cpu_setup(void);
- void mte_suspend_enter(void);
- void mte_suspend_exit(void);
- long set_mte_ctrl(struct task_struct *task, unsigned long arg);
- long get_mte_ctrl(struct task_struct *task);
- int mte_ptrace_copy_tags(struct task_struct *child, long request,
- unsigned long addr, unsigned long data);
- size_t mte_probe_user_range(const char __user *uaddr, size_t size);
- #else /* CONFIG_ARM64_MTE */
- /* unused if !CONFIG_ARM64_MTE, silence the compiler */
- #define PG_mte_tagged 0
- static inline void set_page_mte_tagged(struct page *page)
- {
- }
- static inline bool page_mte_tagged(struct page *page)
- {
- return false;
- }
- static inline void mte_zero_clear_page_tags(void *addr)
- {
- }
- static inline void mte_sync_tags(pte_t pte)
- {
- }
- static inline void mte_copy_page_tags(void *kto, const void *kfrom)
- {
- }
- static inline void mte_thread_init_user(void)
- {
- }
- static inline void mte_thread_switch(struct task_struct *next)
- {
- }
- static inline void mte_suspend_enter(void)
- {
- }
- static inline void mte_suspend_exit(void)
- {
- }
- static inline long set_mte_ctrl(struct task_struct *task, unsigned long arg)
- {
- return 0;
- }
- static inline long get_mte_ctrl(struct task_struct *task)
- {
- return 0;
- }
- static inline int mte_ptrace_copy_tags(struct task_struct *child,
- long request, unsigned long addr,
- unsigned long data)
- {
- return -EIO;
- }
- #endif /* CONFIG_ARM64_MTE */
- static inline void mte_disable_tco_entry(struct task_struct *task)
- {
- if (!system_supports_mte())
- return;
- /*
- * Re-enable tag checking (TCO set on exception entry). This is only
- * necessary if MTE is enabled in either the kernel or the userspace
- * task in synchronous or asymmetric mode (SCTLR_EL1.TCF0 bit 0 is set
- * for both). With MTE disabled in the kernel and disabled or
- * asynchronous in userspace, tag check faults (including in uaccesses)
- * are not reported, therefore there is no need to re-enable checking.
- * This is beneficial on microarchitectures where re-enabling TCO is
- * expensive.
- */
- if (kasan_hw_tags_enabled() ||
- (task->thread.sctlr_user & (1UL << SCTLR_EL1_TCF0_SHIFT)))
- asm volatile(SET_PSTATE_TCO(0));
- }
- #ifdef CONFIG_KASAN_HW_TAGS
- void mte_check_tfsr_el1(void);
- static inline void mte_check_tfsr_entry(void)
- {
- if (!system_supports_mte())
- return;
- mte_check_tfsr_el1();
- }
- static inline void mte_check_tfsr_exit(void)
- {
- if (!system_supports_mte())
- return;
- /*
- * The asynchronous faults are sync'ed automatically with
- * TFSR_EL1 on kernel entry but for exit an explicit dsb()
- * is required.
- */
- dsb(nsh);
- isb();
- mte_check_tfsr_el1();
- }
- #else
- static inline void mte_check_tfsr_el1(void)
- {
- }
- static inline void mte_check_tfsr_entry(void)
- {
- }
- static inline void mte_check_tfsr_exit(void)
- {
- }
- #endif /* CONFIG_KASAN_HW_TAGS */
- #endif /* __ASSEMBLY__ */
- #endif /* __ASM_MTE_H */
|