FROMLIST: mm: adding speculative page fault failure trace events
This patch a set of new trace events to collect the speculative page fault event failures. Signed-off-by: Laurent Dufour <ldufour@linux.vnet.ibm.com> Change-Id: I9dafba293edf40bdad4ae241d105ecdfb42579c1 Link: https://lore.kernel.org/lkml/1523975611-15978-20-git-send-email-ldufour@linux.vnet.ibm.com/ Bug: 161210518 Signed-off-by: Vinayak Menon <vinmenon@codeaurora.org>
This commit is contained in:

committed by
Suren Baghdasaryan

parent
1c53717440
commit
736ae8bde8
89
include/trace/events/pagefault.h
Normal file
89
include/trace/events/pagefault.h
Normal file
@@ -0,0 +1,89 @@
|
||||
/* SPDX-License-Identifier: GPL-2.0 */
|
||||
#undef TRACE_SYSTEM
|
||||
#undef TRACE_INCLUDE_PATH
|
||||
#define TRACE_SYSTEM pagefault
|
||||
|
||||
#if !defined(_TRACE_PAGEFAULT_H) || defined(TRACE_HEADER_MULTI_READ)
|
||||
#define _TRACE_PAGEFAULT_H
|
||||
|
||||
#include <linux/tracepoint.h>
|
||||
#include <linux/mm.h>
|
||||
|
||||
DECLARE_EVENT_CLASS(spf,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field(unsigned long, caller)
|
||||
__field(unsigned long, vm_start)
|
||||
__field(unsigned long, vm_end)
|
||||
__field(unsigned long, address)
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->caller = caller;
|
||||
__entry->vm_start = vma->vm_start;
|
||||
__entry->vm_end = vma->vm_end;
|
||||
__entry->address = address;
|
||||
),
|
||||
|
||||
TP_printk("ip:%lx vma:%lx-%lx address:%lx",
|
||||
__entry->caller, __entry->vm_start, __entry->vm_end,
|
||||
__entry->address)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(spf, spf_pte_lock,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(spf, spf_vma_changed,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(spf, spf_vma_noanon,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(spf, spf_vma_notsup,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(spf, spf_vma_access,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(spf, spf_pmd_changed,
|
||||
|
||||
TP_PROTO(unsigned long caller,
|
||||
struct vm_area_struct *vma, unsigned long address),
|
||||
|
||||
TP_ARGS(caller, vma, address)
|
||||
);
|
||||
|
||||
#endif /* _TRACE_PAGEFAULT_H */
|
||||
|
||||
/* This part must be outside protection */
|
||||
#include <trace/define_trace.h>
|
65
mm/memory.c
65
mm/memory.c
@@ -87,6 +87,9 @@
|
||||
#include "pgalloc-track.h"
|
||||
#include "internal.h"
|
||||
|
||||
#define CREATE_TRACE_POINTS
|
||||
#include <trace/events/pagefault.h>
|
||||
|
||||
#if defined(LAST_CPUPID_NOT_IN_PAGE_FLAGS) && !defined(CONFIG_COMPILE_TEST)
|
||||
#warning Unfortunate NUMA and NUMA Balancing config, growing page-frame for last_cpupid.
|
||||
#endif
|
||||
@@ -2588,8 +2591,10 @@ static bool pte_spinlock(struct vm_fault *vmf)
|
||||
}
|
||||
|
||||
local_irq_disable();
|
||||
if (vma_has_changed(vmf))
|
||||
if (vma_has_changed(vmf)) {
|
||||
trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
||||
/*
|
||||
@@ -2597,16 +2602,21 @@ static bool pte_spinlock(struct vm_fault *vmf)
|
||||
* is not a huge collapse operation in progress in our back.
|
||||
*/
|
||||
pmdval = READ_ONCE(*vmf->pmd);
|
||||
if (!pmd_same(pmdval, vmf->orig_pmd))
|
||||
if (!pmd_same(pmdval, vmf->orig_pmd)) {
|
||||
trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
#endif
|
||||
|
||||
vmf->ptl = pte_lockptr(vmf->vma->vm_mm, vmf->pmd);
|
||||
if (unlikely(!spin_trylock(vmf->ptl)))
|
||||
if (unlikely(!spin_trylock(vmf->ptl))) {
|
||||
trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (vma_has_changed(vmf)) {
|
||||
spin_unlock(vmf->ptl);
|
||||
trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -2639,8 +2649,10 @@ static bool pte_map_lock(struct vm_fault *vmf)
|
||||
* block on the PTL and thus we're safe.
|
||||
*/
|
||||
local_irq_disable();
|
||||
if (vma_has_changed(vmf))
|
||||
if (vma_has_changed(vmf)) {
|
||||
trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
||||
/*
|
||||
@@ -2648,8 +2660,10 @@ static bool pte_map_lock(struct vm_fault *vmf)
|
||||
* is not a huge collapse operation in progress in our back.
|
||||
*/
|
||||
pmdval = READ_ONCE(*vmf->pmd);
|
||||
if (!pmd_same(pmdval, vmf->orig_pmd))
|
||||
if (!pmd_same(pmdval, vmf->orig_pmd)) {
|
||||
trace_spf_pmd_changed(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
@@ -2663,11 +2677,13 @@ static bool pte_map_lock(struct vm_fault *vmf)
|
||||
pte = pte_offset_map(vmf->pmd, vmf->address);
|
||||
if (unlikely(!spin_trylock(ptl))) {
|
||||
pte_unmap(pte);
|
||||
trace_spf_pte_lock(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (vma_has_changed(vmf)) {
|
||||
pte_unmap_unlock(pte, ptl);
|
||||
trace_spf_vma_changed(_RET_IP_, vmf->vma, vmf->address);
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -4827,47 +4843,60 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
|
||||
|
||||
/* rmb <-> seqlock,vma_rb_erase() */
|
||||
seq = raw_read_seqcount(&vma->vm_sequence);
|
||||
if (seq & 1)
|
||||
if (seq & 1) {
|
||||
trace_spf_vma_changed(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
/*
|
||||
* Can't call vm_ops service has we don't know what they would do
|
||||
* with the VMA.
|
||||
* This include huge page from hugetlbfs.
|
||||
*/
|
||||
if (vma->vm_ops)
|
||||
if (vma->vm_ops) {
|
||||
trace_spf_vma_notsup(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
/*
|
||||
* __anon_vma_prepare() requires the mmap_sem to be held
|
||||
* because vm_next and vm_prev must be safe. This can't be guaranteed
|
||||
* in the speculative path.
|
||||
*/
|
||||
if (unlikely(!vma->anon_vma))
|
||||
if (unlikely(!vma->anon_vma)) {
|
||||
trace_spf_vma_notsup(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
vmf.vma_flags = READ_ONCE(vma->vm_flags);
|
||||
vmf.vma_page_prot = READ_ONCE(vma->vm_page_prot);
|
||||
|
||||
/* Can't call userland page fault handler in the speculative path */
|
||||
if (unlikely(vmf.vma_flags & VM_UFFD_MISSING))
|
||||
if (unlikely(vmf.vma_flags & VM_UFFD_MISSING)) {
|
||||
trace_spf_vma_notsup(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP)
|
||||
if (vmf.vma_flags & VM_GROWSDOWN || vmf.vma_flags & VM_GROWSUP) {
|
||||
/*
|
||||
* This could be detected by the check address against VMA's
|
||||
* boundaries but we want to trace it as not supported instead
|
||||
* of changed.
|
||||
*/
|
||||
trace_spf_vma_notsup(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
if (address < READ_ONCE(vma->vm_start)
|
||||
|| READ_ONCE(vma->vm_end) <= address)
|
||||
|| READ_ONCE(vma->vm_end) <= address) {
|
||||
trace_spf_vma_changed(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
if (!arch_vma_access_permitted(vma, flags & FAULT_FLAG_WRITE,
|
||||
flags & FAULT_FLAG_INSTRUCTION,
|
||||
flags & FAULT_FLAG_REMOTE)) {
|
||||
trace_spf_vma_access(_RET_IP_, vma, address);
|
||||
ret = VM_FAULT_SIGSEGV;
|
||||
goto out_put;
|
||||
}
|
||||
@@ -4875,10 +4904,12 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
|
||||
/* This is one is required to check that the VMA has write access set */
|
||||
if (flags & FAULT_FLAG_WRITE) {
|
||||
if (unlikely(!(vmf.vma_flags & VM_WRITE))) {
|
||||
trace_spf_vma_access(_RET_IP_, vma, address);
|
||||
ret = VM_FAULT_SIGSEGV;
|
||||
goto out_put;
|
||||
}
|
||||
} else if (unlikely(!(vmf.vma_flags & (VM_READ|VM_EXEC|VM_WRITE)))) {
|
||||
trace_spf_vma_access(_RET_IP_, vma, address);
|
||||
ret = VM_FAULT_SIGSEGV;
|
||||
goto out_put;
|
||||
}
|
||||
@@ -4894,8 +4925,11 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
|
||||
pol = __get_vma_policy(vma, address);
|
||||
if (!pol)
|
||||
pol = get_task_policy(current);
|
||||
if (pol && pol->mode == MPOL_INTERLEAVE)
|
||||
goto out_put;
|
||||
if (!pol)
|
||||
if (pol && pol->mode == MPOL_INTERLEAVE) {
|
||||
trace_spf_vma_notsup(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
@@ -4968,8 +5002,10 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
|
||||
* We need to re-validate the VMA after checking the bounds, otherwise
|
||||
* we might have a false positive on the bounds.
|
||||
*/
|
||||
if (read_seqcount_retry(&vma->vm_sequence, seq))
|
||||
if (read_seqcount_retry(&vma->vm_sequence, seq)) {
|
||||
trace_spf_vma_changed(_RET_IP_, vma, address);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
mem_cgroup_enter_user_fault();
|
||||
ret = handle_pte_fault(&vmf);
|
||||
@@ -4988,6 +5024,7 @@ int __handle_speculative_fault(struct mm_struct *mm, unsigned long address,
|
||||
return ret;
|
||||
|
||||
out_walk:
|
||||
trace_spf_vma_notsup(_RET_IP_, vma, address);
|
||||
local_irq_enable();
|
||||
out_put:
|
||||
put_vma(vma);
|
||||
|
Reference in New Issue
Block a user