Merge tag 'v5.7-rc1' into locking/kcsan, to resolve conflicts and refresh
Resolve these conflicts: arch/x86/Kconfig arch/x86/kernel/Makefile Do a minor "evil merge" to move the KCSAN entry up a bit by a few lines in the Kconfig to reduce the probability of future conflicts. Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
1
arch/x86/entry/vdso/.gitignore
vendored
1
arch/x86/entry/vdso/.gitignore
vendored
@@ -1,3 +1,4 @@
|
||||
# SPDX-License-Identifier: GPL-2.0-only
|
||||
vdso.lds
|
||||
vdsox32.lds
|
||||
vdso32-syscall-syms.lds
|
||||
|
@@ -57,6 +57,13 @@ SECTIONS
|
||||
*(.gnu.linkonce.b.*)
|
||||
} :text
|
||||
|
||||
/*
|
||||
* Discard .note.gnu.property sections which are unused and have
|
||||
* different alignment requirement from vDSO note sections.
|
||||
*/
|
||||
/DISCARD/ : {
|
||||
*(.note.gnu.property)
|
||||
}
|
||||
.note : { *(.note.*) } :text :note
|
||||
|
||||
.eh_frame_hdr : { *(.eh_frame_hdr) } :text :eh_frame_hdr
|
||||
|
1
arch/x86/entry/vdso/vdso32/.gitignore
vendored
1
arch/x86/entry/vdso/vdso32/.gitignore
vendored
@@ -1 +1,2 @@
|
||||
# SPDX-License-Identifier: GPL-2.0-only
|
||||
vdso32.lds
|
||||
|
@@ -1,10 +1,6 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
#define BUILD_VDSO32
|
||||
|
||||
#ifndef CONFIG_CC_OPTIMIZE_FOR_SIZE
|
||||
#undef CONFIG_OPTIMIZE_INLINING
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_X86_64
|
||||
|
||||
/*
|
||||
@@ -13,6 +9,7 @@
|
||||
*/
|
||||
#undef CONFIG_64BIT
|
||||
#undef CONFIG_X86_64
|
||||
#undef CONFIG_COMPAT
|
||||
#undef CONFIG_PGTABLE_LEVELS
|
||||
#undef CONFIG_ILLEGAL_POINTER_VALUE
|
||||
#undef CONFIG_SPARSEMEM_VMEMMAP
|
||||
|
@@ -38,6 +38,8 @@ struct vdso_data *arch_get_vdso_data(void *vvar_page)
|
||||
}
|
||||
#undef EMIT_VVAR
|
||||
|
||||
unsigned int vclocks_used __read_mostly;
|
||||
|
||||
#if defined(CONFIG_X86_64)
|
||||
unsigned int __read_mostly vdso64_enabled = 1;
|
||||
#endif
|
||||
@@ -219,7 +221,7 @@ static vm_fault_t vvar_fault(const struct vm_special_mapping *sm,
|
||||
} else if (sym_offset == image->sym_pvclock_page) {
|
||||
struct pvclock_vsyscall_time_info *pvti =
|
||||
pvclock_get_pvti_cpu0_va();
|
||||
if (pvti && vclock_was_used(VCLOCK_PVCLOCK)) {
|
||||
if (pvti && vclock_was_used(VDSO_CLOCKMODE_PVCLOCK)) {
|
||||
return vmf_insert_pfn_prot(vma, vmf->address,
|
||||
__pa(pvti) >> PAGE_SHIFT,
|
||||
pgprot_decrypted(vma->vm_page_prot));
|
||||
@@ -227,7 +229,7 @@ static vm_fault_t vvar_fault(const struct vm_special_mapping *sm,
|
||||
} else if (sym_offset == image->sym_hvclock_page) {
|
||||
struct ms_hyperv_tsc_page *tsc_pg = hv_get_tsc_page();
|
||||
|
||||
if (tsc_pg && vclock_was_used(VCLOCK_HVCLOCK))
|
||||
if (tsc_pg && vclock_was_used(VDSO_CLOCKMODE_HVCLOCK))
|
||||
return vmf_insert_pfn(vma, vmf->address,
|
||||
virt_to_phys(tsc_pg) >> PAGE_SHIFT);
|
||||
} else if (sym_offset == image->sym_timens_page) {
|
||||
@@ -445,6 +447,8 @@ __setup("vdso=", vdso_setup);
|
||||
|
||||
static int __init init_vdso(void)
|
||||
{
|
||||
BUILD_BUG_ON(VDSO_CLOCKMODE_MAX >= 32);
|
||||
|
||||
init_vdso_image(&vdso_image_64);
|
||||
|
||||
#ifdef CONFIG_X86_X32_ABI
|
||||
|
Reference in New Issue
Block a user