1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171 |
- // SPDX-License-Identifier: GPL-2.0-only
- /*
- * Handle detection, reporting and mitigation of Spectre v1, v2, v3a and v4, as
- * detailed at:
- *
- * https://developer.arm.com/support/arm-security-updates/speculative-processor-vulnerability
- *
- * This code was originally written hastily under an awful lot of stress and so
- * aspects of it are somewhat hacky. Unfortunately, changing anything in here
- * instantly makes me feel ill. Thanks, Jann. Thann.
- *
- * Copyright (C) 2018 ARM Ltd, All Rights Reserved.
- * Copyright (C) 2020 Google LLC
- *
- * "If there's something strange in your neighbourhood, who you gonna call?"
- *
- * Authors: Will Deacon <[email protected]> and Marc Zyngier <[email protected]>
- */
- #include <linux/arm-smccc.h>
- #include <linux/bpf.h>
- #include <linux/cpu.h>
- #include <linux/device.h>
- #include <linux/nospec.h>
- #include <linux/prctl.h>
- #include <linux/sched/task_stack.h>
- #include <asm/debug-monitors.h>
- #include <asm/insn.h>
- #include <asm/spectre.h>
- #include <asm/traps.h>
- #include <asm/vectors.h>
- #include <asm/virt.h>
- /*
- * We try to ensure that the mitigation state can never change as the result of
- * onlining a late CPU.
- */
- static void update_mitigation_state(enum mitigation_state *oldp,
- enum mitigation_state new)
- {
- enum mitigation_state state;
- do {
- state = READ_ONCE(*oldp);
- if (new <= state)
- break;
- /* Userspace almost certainly can't deal with this. */
- if (WARN_ON(system_capabilities_finalized()))
- break;
- } while (cmpxchg_relaxed(oldp, state, new) != state);
- }
- /*
- * Spectre v1.
- *
- * The kernel can't protect userspace for this one: it's each person for
- * themselves. Advertise what we're doing and be done with it.
- */
- ssize_t cpu_show_spectre_v1(struct device *dev, struct device_attribute *attr,
- char *buf)
- {
- return sprintf(buf, "Mitigation: __user pointer sanitization\n");
- }
- /*
- * Spectre v2.
- *
- * This one sucks. A CPU is either:
- *
- * - Mitigated in hardware and advertised by ID_AA64PFR0_EL1.CSV2.
- * - Mitigated in hardware and listed in our "safe list".
- * - Mitigated in software by firmware.
- * - Mitigated in software by a CPU-specific dance in the kernel and a
- * firmware call at EL2.
- * - Vulnerable.
- *
- * It's not unlikely for different CPUs in a big.LITTLE system to fall into
- * different camps.
- */
- static enum mitigation_state spectre_v2_state;
- static bool __read_mostly __nospectre_v2;
- static int __init parse_spectre_v2_param(char *str)
- {
- __nospectre_v2 = true;
- return 0;
- }
- early_param("nospectre_v2", parse_spectre_v2_param);
- static bool spectre_v2_mitigations_off(void)
- {
- bool ret = __nospectre_v2 || cpu_mitigations_off();
- if (ret)
- pr_info_once("spectre-v2 mitigation disabled by command line option\n");
- return ret;
- }
- static const char *get_bhb_affected_string(enum mitigation_state bhb_state)
- {
- switch (bhb_state) {
- case SPECTRE_UNAFFECTED:
- return "";
- default:
- case SPECTRE_VULNERABLE:
- return ", but not BHB";
- case SPECTRE_MITIGATED:
- return ", BHB";
- }
- }
- static bool _unprivileged_ebpf_enabled(void)
- {
- #ifdef CONFIG_BPF_SYSCALL
- return !sysctl_unprivileged_bpf_disabled;
- #else
- return false;
- #endif
- }
- ssize_t cpu_show_spectre_v2(struct device *dev, struct device_attribute *attr,
- char *buf)
- {
- enum mitigation_state bhb_state = arm64_get_spectre_bhb_state();
- const char *bhb_str = get_bhb_affected_string(bhb_state);
- const char *v2_str = "Branch predictor hardening";
- switch (spectre_v2_state) {
- case SPECTRE_UNAFFECTED:
- if (bhb_state == SPECTRE_UNAFFECTED)
- return sprintf(buf, "Not affected\n");
- /*
- * Platforms affected by Spectre-BHB can't report
- * "Not affected" for Spectre-v2.
- */
- v2_str = "CSV2";
- fallthrough;
- case SPECTRE_MITIGATED:
- if (bhb_state == SPECTRE_MITIGATED && _unprivileged_ebpf_enabled())
- return sprintf(buf, "Vulnerable: Unprivileged eBPF enabled\n");
- return sprintf(buf, "Mitigation: %s%s\n", v2_str, bhb_str);
- case SPECTRE_VULNERABLE:
- fallthrough;
- default:
- return sprintf(buf, "Vulnerable\n");
- }
- }
- static enum mitigation_state spectre_v2_get_cpu_hw_mitigation_state(void)
- {
- u64 pfr0;
- static const struct midr_range spectre_v2_safe_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A35),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A53),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
- MIDR_ALL_VERSIONS(MIDR_BRAHMA_B53),
- MIDR_ALL_VERSIONS(MIDR_HISI_TSV110),
- MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_2XX_SILVER),
- MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_3XX_SILVER),
- MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_4XX_SILVER),
- { /* sentinel */ }
- };
- /* If the CPU has CSV2 set, we're safe */
- pfr0 = read_cpuid(ID_AA64PFR0_EL1);
- if (cpuid_feature_extract_unsigned_field(pfr0, ID_AA64PFR0_EL1_CSV2_SHIFT))
- return SPECTRE_UNAFFECTED;
- /* Alternatively, we have a list of unaffected CPUs */
- if (is_midr_in_range_list(read_cpuid_id(), spectre_v2_safe_list))
- return SPECTRE_UNAFFECTED;
- return SPECTRE_VULNERABLE;
- }
- static enum mitigation_state spectre_v2_get_cpu_fw_mitigation_state(void)
- {
- int ret;
- struct arm_smccc_res res;
- arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
- ARM_SMCCC_ARCH_WORKAROUND_1, &res);
- ret = res.a0;
- switch (ret) {
- case SMCCC_RET_SUCCESS:
- return SPECTRE_MITIGATED;
- case SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED:
- return SPECTRE_UNAFFECTED;
- default:
- fallthrough;
- case SMCCC_RET_NOT_SUPPORTED:
- return SPECTRE_VULNERABLE;
- }
- }
- bool has_spectre_v2(const struct arm64_cpu_capabilities *entry, int scope)
- {
- WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
- if (spectre_v2_get_cpu_hw_mitigation_state() == SPECTRE_UNAFFECTED)
- return false;
- if (spectre_v2_get_cpu_fw_mitigation_state() == SPECTRE_UNAFFECTED)
- return false;
- return true;
- }
- enum mitigation_state arm64_get_spectre_v2_state(void)
- {
- return spectre_v2_state;
- }
- DEFINE_PER_CPU_READ_MOSTLY(struct bp_hardening_data, bp_hardening_data);
- static void install_bp_hardening_cb(bp_hardening_cb_t fn)
- {
- __this_cpu_write(bp_hardening_data.fn, fn);
- /*
- * Vinz Clortho takes the hyp_vecs start/end "keys" at
- * the door when we're a guest. Skip the hyp-vectors work.
- */
- if (!is_hyp_mode_available())
- return;
- __this_cpu_write(bp_hardening_data.slot, HYP_VECTOR_SPECTRE_DIRECT);
- }
- /* Called during entry so must be noinstr */
- static noinstr void call_smc_arch_workaround_1(void)
- {
- arm_smccc_1_1_smc(ARM_SMCCC_ARCH_WORKAROUND_1, NULL);
- }
- /* Called during entry so must be noinstr */
- static noinstr void call_hvc_arch_workaround_1(void)
- {
- arm_smccc_1_1_hvc(ARM_SMCCC_ARCH_WORKAROUND_1, NULL);
- }
- /* Called during entry so must be noinstr */
- static noinstr void qcom_link_stack_sanitisation(void)
- {
- u64 tmp;
- asm volatile("mov %0, x30 \n"
- ".rept 16 \n"
- "bl . + 4 \n"
- ".endr \n"
- "mov x30, %0 \n"
- : "=&r" (tmp));
- }
- static bp_hardening_cb_t spectre_v2_get_sw_mitigation_cb(void)
- {
- u32 midr = read_cpuid_id();
- if (((midr & MIDR_CPU_MODEL_MASK) != MIDR_QCOM_FALKOR) &&
- ((midr & MIDR_CPU_MODEL_MASK) != MIDR_QCOM_FALKOR_V1))
- return NULL;
- return qcom_link_stack_sanitisation;
- }
- static enum mitigation_state spectre_v2_enable_fw_mitigation(void)
- {
- bp_hardening_cb_t cb;
- enum mitigation_state state;
- state = spectre_v2_get_cpu_fw_mitigation_state();
- if (state != SPECTRE_MITIGATED)
- return state;
- if (spectre_v2_mitigations_off())
- return SPECTRE_VULNERABLE;
- switch (arm_smccc_1_1_get_conduit()) {
- case SMCCC_CONDUIT_HVC:
- cb = call_hvc_arch_workaround_1;
- break;
- case SMCCC_CONDUIT_SMC:
- cb = call_smc_arch_workaround_1;
- break;
- default:
- return SPECTRE_VULNERABLE;
- }
- /*
- * Prefer a CPU-specific workaround if it exists. Note that we
- * still rely on firmware for the mitigation at EL2.
- */
- cb = spectre_v2_get_sw_mitigation_cb() ?: cb;
- install_bp_hardening_cb(cb);
- return SPECTRE_MITIGATED;
- }
- void spectre_v2_enable_mitigation(const struct arm64_cpu_capabilities *__unused)
- {
- enum mitigation_state state;
- WARN_ON(preemptible());
- state = spectre_v2_get_cpu_hw_mitigation_state();
- if (state == SPECTRE_VULNERABLE)
- state = spectre_v2_enable_fw_mitigation();
- update_mitigation_state(&spectre_v2_state, state);
- }
- /*
- * Spectre-v3a.
- *
- * Phew, there's not an awful lot to do here! We just instruct EL2 to use
- * an indirect trampoline for the hyp vectors so that guests can't read
- * VBAR_EL2 to defeat randomisation of the hypervisor VA layout.
- */
- bool has_spectre_v3a(const struct arm64_cpu_capabilities *entry, int scope)
- {
- static const struct midr_range spectre_v3a_unsafe_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A57),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
- {},
- };
- WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
- return is_midr_in_range_list(read_cpuid_id(), spectre_v3a_unsafe_list);
- }
- void spectre_v3a_enable_mitigation(const struct arm64_cpu_capabilities *__unused)
- {
- struct bp_hardening_data *data = this_cpu_ptr(&bp_hardening_data);
- if (this_cpu_has_cap(ARM64_SPECTRE_V3A))
- data->slot += HYP_VECTOR_INDIRECT;
- }
- /*
- * Spectre v4.
- *
- * If you thought Spectre v2 was nasty, wait until you see this mess. A CPU is
- * either:
- *
- * - Mitigated in hardware and listed in our "safe list".
- * - Mitigated in hardware via PSTATE.SSBS.
- * - Mitigated in software by firmware (sometimes referred to as SSBD).
- *
- * Wait, that doesn't sound so bad, does it? Keep reading...
- *
- * A major source of headaches is that the software mitigation is enabled both
- * on a per-task basis, but can also be forced on for the kernel, necessitating
- * both context-switch *and* entry/exit hooks. To make it even worse, some CPUs
- * allow EL0 to toggle SSBS directly, which can end up with the prctl() state
- * being stale when re-entering the kernel. The usual big.LITTLE caveats apply,
- * so you can have systems that have both firmware and SSBS mitigations. This
- * means we actually have to reject late onlining of CPUs with mitigations if
- * all of the currently onlined CPUs are safelisted, as the mitigation tends to
- * be opt-in for userspace. Yes, really, the cure is worse than the disease.
- *
- * The only good part is that if the firmware mitigation is present, then it is
- * present for all CPUs, meaning we don't have to worry about late onlining of a
- * vulnerable CPU if one of the boot CPUs is using the firmware mitigation.
- *
- * Give me a VAX-11/780 any day of the week...
- */
- static enum mitigation_state spectre_v4_state;
- /* This is the per-cpu state tracking whether we need to talk to firmware */
- DEFINE_PER_CPU_READ_MOSTLY(u64, arm64_ssbd_callback_required);
- enum spectre_v4_policy {
- SPECTRE_V4_POLICY_MITIGATION_DYNAMIC,
- SPECTRE_V4_POLICY_MITIGATION_ENABLED,
- SPECTRE_V4_POLICY_MITIGATION_DISABLED,
- };
- static enum spectre_v4_policy __read_mostly __spectre_v4_policy;
- static const struct spectre_v4_param {
- const char *str;
- enum spectre_v4_policy policy;
- } spectre_v4_params[] = {
- { "force-on", SPECTRE_V4_POLICY_MITIGATION_ENABLED, },
- { "force-off", SPECTRE_V4_POLICY_MITIGATION_DISABLED, },
- { "kernel", SPECTRE_V4_POLICY_MITIGATION_DYNAMIC, },
- };
- static int __init parse_spectre_v4_param(char *str)
- {
- int i;
- if (!str || !str[0])
- return -EINVAL;
- for (i = 0; i < ARRAY_SIZE(spectre_v4_params); i++) {
- const struct spectre_v4_param *param = &spectre_v4_params[i];
- if (strncmp(str, param->str, strlen(param->str)))
- continue;
- __spectre_v4_policy = param->policy;
- return 0;
- }
- return -EINVAL;
- }
- early_param("ssbd", parse_spectre_v4_param);
- /*
- * Because this was all written in a rush by people working in different silos,
- * we've ended up with multiple command line options to control the same thing.
- * Wrap these up in some helpers, which prefer disabling the mitigation if faced
- * with contradictory parameters. The mitigation is always either "off",
- * "dynamic" or "on".
- */
- static bool spectre_v4_mitigations_off(void)
- {
- bool ret = cpu_mitigations_off() ||
- __spectre_v4_policy == SPECTRE_V4_POLICY_MITIGATION_DISABLED;
- if (ret)
- pr_info_once("spectre-v4 mitigation disabled by command-line option\n");
- return ret;
- }
- /* Do we need to toggle the mitigation state on entry to/exit from the kernel? */
- static bool spectre_v4_mitigations_dynamic(void)
- {
- return !spectre_v4_mitigations_off() &&
- __spectre_v4_policy == SPECTRE_V4_POLICY_MITIGATION_DYNAMIC;
- }
- static bool spectre_v4_mitigations_on(void)
- {
- return !spectre_v4_mitigations_off() &&
- __spectre_v4_policy == SPECTRE_V4_POLICY_MITIGATION_ENABLED;
- }
- ssize_t cpu_show_spec_store_bypass(struct device *dev,
- struct device_attribute *attr, char *buf)
- {
- switch (spectre_v4_state) {
- case SPECTRE_UNAFFECTED:
- return sprintf(buf, "Not affected\n");
- case SPECTRE_MITIGATED:
- return sprintf(buf, "Mitigation: Speculative Store Bypass disabled via prctl\n");
- case SPECTRE_VULNERABLE:
- fallthrough;
- default:
- return sprintf(buf, "Vulnerable\n");
- }
- }
- enum mitigation_state arm64_get_spectre_v4_state(void)
- {
- return spectre_v4_state;
- }
- static enum mitigation_state spectre_v4_get_cpu_hw_mitigation_state(void)
- {
- static const struct midr_range spectre_v4_safe_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A35),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A53),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
- MIDR_ALL_VERSIONS(MIDR_BRAHMA_B53),
- MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_3XX_SILVER),
- MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_4XX_SILVER),
- { /* sentinel */ },
- };
- if (is_midr_in_range_list(read_cpuid_id(), spectre_v4_safe_list))
- return SPECTRE_UNAFFECTED;
- /* CPU features are detected first */
- if (this_cpu_has_cap(ARM64_SSBS))
- return SPECTRE_MITIGATED;
- return SPECTRE_VULNERABLE;
- }
- static enum mitigation_state spectre_v4_get_cpu_fw_mitigation_state(void)
- {
- int ret;
- struct arm_smccc_res res;
- arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
- ARM_SMCCC_ARCH_WORKAROUND_2, &res);
- ret = res.a0;
- switch (ret) {
- case SMCCC_RET_SUCCESS:
- return SPECTRE_MITIGATED;
- case SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED:
- fallthrough;
- case SMCCC_RET_NOT_REQUIRED:
- return SPECTRE_UNAFFECTED;
- default:
- fallthrough;
- case SMCCC_RET_NOT_SUPPORTED:
- return SPECTRE_VULNERABLE;
- }
- }
- bool has_spectre_v4(const struct arm64_cpu_capabilities *cap, int scope)
- {
- enum mitigation_state state;
- WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
- state = spectre_v4_get_cpu_hw_mitigation_state();
- if (state == SPECTRE_VULNERABLE)
- state = spectre_v4_get_cpu_fw_mitigation_state();
- return state != SPECTRE_UNAFFECTED;
- }
- static int ssbs_emulation_handler(struct pt_regs *regs, u32 instr)
- {
- if (user_mode(regs))
- return 1;
- if (instr & BIT(PSTATE_Imm_shift))
- regs->pstate |= PSR_SSBS_BIT;
- else
- regs->pstate &= ~PSR_SSBS_BIT;
- arm64_skip_faulting_instruction(regs, 4);
- return 0;
- }
- static struct undef_hook ssbs_emulation_hook = {
- .instr_mask = ~(1U << PSTATE_Imm_shift),
- .instr_val = 0xd500401f | PSTATE_SSBS,
- .fn = ssbs_emulation_handler,
- };
- static enum mitigation_state spectre_v4_enable_hw_mitigation(void)
- {
- static bool undef_hook_registered = false;
- static DEFINE_RAW_SPINLOCK(hook_lock);
- enum mitigation_state state;
- /*
- * If the system is mitigated but this CPU doesn't have SSBS, then
- * we must be on the safelist and there's nothing more to do.
- */
- state = spectre_v4_get_cpu_hw_mitigation_state();
- if (state != SPECTRE_MITIGATED || !this_cpu_has_cap(ARM64_SSBS))
- return state;
- raw_spin_lock(&hook_lock);
- if (!undef_hook_registered) {
- register_undef_hook(&ssbs_emulation_hook);
- undef_hook_registered = true;
- }
- raw_spin_unlock(&hook_lock);
- if (spectre_v4_mitigations_off()) {
- sysreg_clear_set(sctlr_el1, 0, SCTLR_ELx_DSSBS);
- set_pstate_ssbs(1);
- return SPECTRE_VULNERABLE;
- }
- /* SCTLR_EL1.DSSBS was initialised to 0 during boot */
- set_pstate_ssbs(0);
- return SPECTRE_MITIGATED;
- }
- /*
- * Patch a branch over the Spectre-v4 mitigation code with a NOP so that
- * we fallthrough and check whether firmware needs to be called on this CPU.
- */
- void __init spectre_v4_patch_fw_mitigation_enable(struct alt_instr *alt,
- __le32 *origptr,
- __le32 *updptr, int nr_inst)
- {
- BUG_ON(nr_inst != 1); /* Branch -> NOP */
- if (spectre_v4_mitigations_off())
- return;
- if (cpus_have_cap(ARM64_SSBS))
- return;
- if (spectre_v4_mitigations_dynamic())
- *updptr = cpu_to_le32(aarch64_insn_gen_nop());
- }
- /*
- * Patch a NOP in the Spectre-v4 mitigation code with an SMC/HVC instruction
- * to call into firmware to adjust the mitigation state.
- */
- void __init smccc_patch_fw_mitigation_conduit(struct alt_instr *alt,
- __le32 *origptr,
- __le32 *updptr, int nr_inst)
- {
- u32 insn;
- BUG_ON(nr_inst != 1); /* NOP -> HVC/SMC */
- switch (arm_smccc_1_1_get_conduit()) {
- case SMCCC_CONDUIT_HVC:
- insn = aarch64_insn_get_hvc_value();
- break;
- case SMCCC_CONDUIT_SMC:
- insn = aarch64_insn_get_smc_value();
- break;
- default:
- return;
- }
- *updptr = cpu_to_le32(insn);
- }
- static enum mitigation_state spectre_v4_enable_fw_mitigation(void)
- {
- enum mitigation_state state;
- state = spectre_v4_get_cpu_fw_mitigation_state();
- if (state != SPECTRE_MITIGATED)
- return state;
- if (spectre_v4_mitigations_off()) {
- arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_WORKAROUND_2, false, NULL);
- return SPECTRE_VULNERABLE;
- }
- arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_WORKAROUND_2, true, NULL);
- if (spectre_v4_mitigations_dynamic())
- __this_cpu_write(arm64_ssbd_callback_required, 1);
- return SPECTRE_MITIGATED;
- }
- void spectre_v4_enable_mitigation(const struct arm64_cpu_capabilities *__unused)
- {
- enum mitigation_state state;
- WARN_ON(preemptible());
- state = spectre_v4_enable_hw_mitigation();
- if (state == SPECTRE_VULNERABLE)
- state = spectre_v4_enable_fw_mitigation();
- update_mitigation_state(&spectre_v4_state, state);
- }
- static void __update_pstate_ssbs(struct pt_regs *regs, bool state)
- {
- u64 bit = compat_user_mode(regs) ? PSR_AA32_SSBS_BIT : PSR_SSBS_BIT;
- if (state)
- regs->pstate |= bit;
- else
- regs->pstate &= ~bit;
- }
- void spectre_v4_enable_task_mitigation(struct task_struct *tsk)
- {
- struct pt_regs *regs = task_pt_regs(tsk);
- bool ssbs = false, kthread = tsk->flags & PF_KTHREAD;
- if (spectre_v4_mitigations_off())
- ssbs = true;
- else if (spectre_v4_mitigations_dynamic() && !kthread)
- ssbs = !test_tsk_thread_flag(tsk, TIF_SSBD);
- __update_pstate_ssbs(regs, ssbs);
- }
- /*
- * The Spectre-v4 mitigation can be controlled via a prctl() from userspace.
- * This is interesting because the "speculation disabled" behaviour can be
- * configured so that it is preserved across exec(), which means that the
- * prctl() may be necessary even when PSTATE.SSBS can be toggled directly
- * from userspace.
- */
- static void ssbd_prctl_enable_mitigation(struct task_struct *task)
- {
- task_clear_spec_ssb_noexec(task);
- task_set_spec_ssb_disable(task);
- set_tsk_thread_flag(task, TIF_SSBD);
- }
- static void ssbd_prctl_disable_mitigation(struct task_struct *task)
- {
- task_clear_spec_ssb_noexec(task);
- task_clear_spec_ssb_disable(task);
- clear_tsk_thread_flag(task, TIF_SSBD);
- }
- static int ssbd_prctl_set(struct task_struct *task, unsigned long ctrl)
- {
- switch (ctrl) {
- case PR_SPEC_ENABLE:
- /* Enable speculation: disable mitigation */
- /*
- * Force disabled speculation prevents it from being
- * re-enabled.
- */
- if (task_spec_ssb_force_disable(task))
- return -EPERM;
- /*
- * If the mitigation is forced on, then speculation is forced
- * off and we again prevent it from being re-enabled.
- */
- if (spectre_v4_mitigations_on())
- return -EPERM;
- ssbd_prctl_disable_mitigation(task);
- break;
- case PR_SPEC_FORCE_DISABLE:
- /* Force disable speculation: force enable mitigation */
- /*
- * If the mitigation is forced off, then speculation is forced
- * on and we prevent it from being disabled.
- */
- if (spectre_v4_mitigations_off())
- return -EPERM;
- task_set_spec_ssb_force_disable(task);
- fallthrough;
- case PR_SPEC_DISABLE:
- /* Disable speculation: enable mitigation */
- /* Same as PR_SPEC_FORCE_DISABLE */
- if (spectre_v4_mitigations_off())
- return -EPERM;
- ssbd_prctl_enable_mitigation(task);
- break;
- case PR_SPEC_DISABLE_NOEXEC:
- /* Disable speculation until execve(): enable mitigation */
- /*
- * If the mitigation state is forced one way or the other, then
- * we must fail now before we try to toggle it on execve().
- */
- if (task_spec_ssb_force_disable(task) ||
- spectre_v4_mitigations_off() ||
- spectre_v4_mitigations_on()) {
- return -EPERM;
- }
- ssbd_prctl_enable_mitigation(task);
- task_set_spec_ssb_noexec(task);
- break;
- default:
- return -ERANGE;
- }
- spectre_v4_enable_task_mitigation(task);
- return 0;
- }
- int arch_prctl_spec_ctrl_set(struct task_struct *task, unsigned long which,
- unsigned long ctrl)
- {
- switch (which) {
- case PR_SPEC_STORE_BYPASS:
- return ssbd_prctl_set(task, ctrl);
- default:
- return -ENODEV;
- }
- }
- static int ssbd_prctl_get(struct task_struct *task)
- {
- switch (spectre_v4_state) {
- case SPECTRE_UNAFFECTED:
- return PR_SPEC_NOT_AFFECTED;
- case SPECTRE_MITIGATED:
- if (spectre_v4_mitigations_on())
- return PR_SPEC_NOT_AFFECTED;
- if (spectre_v4_mitigations_dynamic())
- break;
- /* Mitigations are disabled, so we're vulnerable. */
- fallthrough;
- case SPECTRE_VULNERABLE:
- fallthrough;
- default:
- return PR_SPEC_ENABLE;
- }
- /* Check the mitigation state for this task */
- if (task_spec_ssb_force_disable(task))
- return PR_SPEC_PRCTL | PR_SPEC_FORCE_DISABLE;
- if (task_spec_ssb_noexec(task))
- return PR_SPEC_PRCTL | PR_SPEC_DISABLE_NOEXEC;
- if (task_spec_ssb_disable(task))
- return PR_SPEC_PRCTL | PR_SPEC_DISABLE;
- return PR_SPEC_PRCTL | PR_SPEC_ENABLE;
- }
- int arch_prctl_spec_ctrl_get(struct task_struct *task, unsigned long which)
- {
- switch (which) {
- case PR_SPEC_STORE_BYPASS:
- return ssbd_prctl_get(task);
- default:
- return -ENODEV;
- }
- }
- /*
- * Spectre BHB.
- *
- * A CPU is either:
- * - Mitigated by a branchy loop a CPU specific number of times, and listed
- * in our "loop mitigated list".
- * - Mitigated in software by the firmware Spectre v2 call.
- * - Has the ClearBHB instruction to perform the mitigation.
- * - Has the 'Exception Clears Branch History Buffer' (ECBHB) feature, so no
- * software mitigation in the vectors is needed.
- * - Has CSV2.3, so is unaffected.
- */
- static enum mitigation_state spectre_bhb_state;
- enum mitigation_state arm64_get_spectre_bhb_state(void)
- {
- return spectre_bhb_state;
- }
- enum bhb_mitigation_bits {
- BHB_LOOP,
- BHB_FW,
- BHB_HW,
- BHB_INSN,
- };
- static unsigned long system_bhb_mitigations;
- /*
- * This must be called with SCOPE_LOCAL_CPU for each type of CPU, before any
- * SCOPE_SYSTEM call will give the right answer.
- */
- u8 spectre_bhb_loop_affected(int scope)
- {
- u8 k = 0;
- static u8 max_bhb_k;
- if (scope == SCOPE_LOCAL_CPU) {
- static const struct midr_range spectre_bhb_k32_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A78),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A78AE),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A78C),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_X1),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A710),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_X2),
- MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N2),
- MIDR_ALL_VERSIONS(MIDR_NEOVERSE_V1),
- {},
- };
- static const struct midr_range spectre_bhb_k24_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A76),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A77),
- MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N1),
- {},
- };
- static const struct midr_range spectre_bhb_k11_list[] = {
- MIDR_ALL_VERSIONS(MIDR_AMPERE1),
- {},
- };
- static const struct midr_range spectre_bhb_k8_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A57),
- {},
- };
- if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k32_list))
- k = 32;
- else if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k24_list))
- k = 24;
- else if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k11_list))
- k = 11;
- else if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k8_list))
- k = 8;
- max_bhb_k = max(max_bhb_k, k);
- } else {
- k = max_bhb_k;
- }
- return k;
- }
- static enum mitigation_state spectre_bhb_get_cpu_fw_mitigation_state(void)
- {
- int ret;
- struct arm_smccc_res res;
- arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
- ARM_SMCCC_ARCH_WORKAROUND_3, &res);
- ret = res.a0;
- switch (ret) {
- case SMCCC_RET_SUCCESS:
- return SPECTRE_MITIGATED;
- case SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED:
- return SPECTRE_UNAFFECTED;
- default:
- fallthrough;
- case SMCCC_RET_NOT_SUPPORTED:
- return SPECTRE_VULNERABLE;
- }
- }
- static bool is_spectre_bhb_fw_affected(int scope)
- {
- static bool system_affected;
- enum mitigation_state fw_state;
- bool has_smccc = arm_smccc_1_1_get_conduit() != SMCCC_CONDUIT_NONE;
- static const struct midr_range spectre_bhb_firmware_mitigated_list[] = {
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A73),
- MIDR_ALL_VERSIONS(MIDR_CORTEX_A75),
- {},
- };
- bool cpu_in_list = is_midr_in_range_list(read_cpuid_id(),
- spectre_bhb_firmware_mitigated_list);
- if (scope != SCOPE_LOCAL_CPU)
- return system_affected;
- fw_state = spectre_bhb_get_cpu_fw_mitigation_state();
- if (cpu_in_list || (has_smccc && fw_state == SPECTRE_MITIGATED)) {
- system_affected = true;
- return true;
- }
- return false;
- }
- static bool supports_ecbhb(int scope)
- {
- u64 mmfr1;
- if (scope == SCOPE_LOCAL_CPU)
- mmfr1 = read_sysreg_s(SYS_ID_AA64MMFR1_EL1);
- else
- mmfr1 = read_sanitised_ftr_reg(SYS_ID_AA64MMFR1_EL1);
- return cpuid_feature_extract_unsigned_field(mmfr1,
- ID_AA64MMFR1_EL1_ECBHB_SHIFT);
- }
- bool is_spectre_bhb_affected(const struct arm64_cpu_capabilities *entry,
- int scope)
- {
- WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
- if (supports_csv2p3(scope))
- return false;
- if (supports_clearbhb(scope))
- return true;
- if (spectre_bhb_loop_affected(scope))
- return true;
- if (is_spectre_bhb_fw_affected(scope))
- return true;
- return false;
- }
- static void this_cpu_set_vectors(enum arm64_bp_harden_el1_vectors slot)
- {
- const char *v = arm64_get_bp_hardening_vector(slot);
- if (slot < 0)
- return;
- __this_cpu_write(this_cpu_vector, v);
- /*
- * When KPTI is in use, the vectors are switched when exiting to
- * user-space.
- */
- if (arm64_kernel_unmapped_at_el0())
- return;
- write_sysreg(v, vbar_el1);
- isb();
- }
- static bool __read_mostly __nospectre_bhb;
- static int __init parse_spectre_bhb_param(char *str)
- {
- __nospectre_bhb = true;
- return 0;
- }
- early_param("nospectre_bhb", parse_spectre_bhb_param);
- void spectre_bhb_enable_mitigation(const struct arm64_cpu_capabilities *entry)
- {
- bp_hardening_cb_t cpu_cb;
- enum mitigation_state fw_state, state = SPECTRE_VULNERABLE;
- struct bp_hardening_data *data = this_cpu_ptr(&bp_hardening_data);
- if (!is_spectre_bhb_affected(entry, SCOPE_LOCAL_CPU))
- return;
- if (arm64_get_spectre_v2_state() == SPECTRE_VULNERABLE) {
- /* No point mitigating Spectre-BHB alone. */
- } else if (!IS_ENABLED(CONFIG_MITIGATE_SPECTRE_BRANCH_HISTORY)) {
- pr_info_once("spectre-bhb mitigation disabled by compile time option\n");
- } else if (cpu_mitigations_off() || __nospectre_bhb) {
- pr_info_once("spectre-bhb mitigation disabled by command line option\n");
- } else if (supports_ecbhb(SCOPE_LOCAL_CPU)) {
- state = SPECTRE_MITIGATED;
- set_bit(BHB_HW, &system_bhb_mitigations);
- } else if (supports_clearbhb(SCOPE_LOCAL_CPU)) {
- /*
- * Ensure KVM uses the indirect vector which will have ClearBHB
- * added.
- */
- if (!data->slot)
- data->slot = HYP_VECTOR_INDIRECT;
- this_cpu_set_vectors(EL1_VECTOR_BHB_CLEAR_INSN);
- state = SPECTRE_MITIGATED;
- set_bit(BHB_INSN, &system_bhb_mitigations);
- } else if (spectre_bhb_loop_affected(SCOPE_LOCAL_CPU)) {
- /*
- * Ensure KVM uses the indirect vector which will have the
- * branchy-loop added. A57/A72-r0 will already have selected
- * the spectre-indirect vector, which is sufficient for BHB
- * too.
- */
- if (!data->slot)
- data->slot = HYP_VECTOR_INDIRECT;
- this_cpu_set_vectors(EL1_VECTOR_BHB_LOOP);
- state = SPECTRE_MITIGATED;
- set_bit(BHB_LOOP, &system_bhb_mitigations);
- } else if (is_spectre_bhb_fw_affected(SCOPE_LOCAL_CPU)) {
- fw_state = spectre_bhb_get_cpu_fw_mitigation_state();
- if (fw_state == SPECTRE_MITIGATED) {
- /*
- * Ensure KVM uses one of the spectre bp_hardening
- * vectors. The indirect vector doesn't include the EL3
- * call, so needs upgrading to
- * HYP_VECTOR_SPECTRE_INDIRECT.
- */
- if (!data->slot || data->slot == HYP_VECTOR_INDIRECT)
- data->slot += 1;
- this_cpu_set_vectors(EL1_VECTOR_BHB_FW);
- /*
- * The WA3 call in the vectors supersedes the WA1 call
- * made during context-switch. Uninstall any firmware
- * bp_hardening callback.
- */
- cpu_cb = spectre_v2_get_sw_mitigation_cb();
- if (__this_cpu_read(bp_hardening_data.fn) != cpu_cb)
- __this_cpu_write(bp_hardening_data.fn, NULL);
- state = SPECTRE_MITIGATED;
- set_bit(BHB_FW, &system_bhb_mitigations);
- }
- }
- update_mitigation_state(&spectre_bhb_state, state);
- }
- /* Patched to NOP when enabled */
- void noinstr spectre_bhb_patch_loop_mitigation_enable(struct alt_instr *alt,
- __le32 *origptr,
- __le32 *updptr, int nr_inst)
- {
- BUG_ON(nr_inst != 1);
- if (test_bit(BHB_LOOP, &system_bhb_mitigations))
- *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
- }
- /* Patched to NOP when enabled */
- void noinstr spectre_bhb_patch_fw_mitigation_enabled(struct alt_instr *alt,
- __le32 *origptr,
- __le32 *updptr, int nr_inst)
- {
- BUG_ON(nr_inst != 1);
- if (test_bit(BHB_FW, &system_bhb_mitigations))
- *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
- }
- /* Patched to correct the immediate */
- void noinstr spectre_bhb_patch_loop_iter(struct alt_instr *alt,
- __le32 *origptr, __le32 *updptr, int nr_inst)
- {
- u8 rd;
- u32 insn;
- u16 loop_count = spectre_bhb_loop_affected(SCOPE_SYSTEM);
- BUG_ON(nr_inst != 1); /* MOV -> MOV */
- if (!IS_ENABLED(CONFIG_MITIGATE_SPECTRE_BRANCH_HISTORY))
- return;
- insn = le32_to_cpu(*origptr);
- rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, insn);
- insn = aarch64_insn_gen_movewide(rd, loop_count, 0,
- AARCH64_INSN_VARIANT_64BIT,
- AARCH64_INSN_MOVEWIDE_ZERO);
- *updptr++ = cpu_to_le32(insn);
- }
- /* Patched to mov WA3 when supported */
- void noinstr spectre_bhb_patch_wa3(struct alt_instr *alt,
- __le32 *origptr, __le32 *updptr, int nr_inst)
- {
- u8 rd;
- u32 insn;
- BUG_ON(nr_inst != 1); /* MOV -> MOV */
- if (!IS_ENABLED(CONFIG_MITIGATE_SPECTRE_BRANCH_HISTORY) ||
- !test_bit(BHB_FW, &system_bhb_mitigations))
- return;
- insn = le32_to_cpu(*origptr);
- rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, insn);
- insn = aarch64_insn_gen_logical_immediate(AARCH64_INSN_LOGIC_ORR,
- AARCH64_INSN_VARIANT_32BIT,
- AARCH64_INSN_REG_ZR, rd,
- ARM_SMCCC_ARCH_WORKAROUND_3);
- if (WARN_ON_ONCE(insn == AARCH64_BREAK_FAULT))
- return;
- *updptr++ = cpu_to_le32(insn);
- }
- /* Patched to NOP when not supported */
- void __init spectre_bhb_patch_clearbhb(struct alt_instr *alt,
- __le32 *origptr, __le32 *updptr, int nr_inst)
- {
- BUG_ON(nr_inst != 2);
- if (test_bit(BHB_INSN, &system_bhb_mitigations))
- return;
- *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
- *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
- }
- #ifdef CONFIG_BPF_SYSCALL
- #define EBPF_WARN "Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks!\n"
- void unpriv_ebpf_notify(int new_state)
- {
- if (spectre_v2_state == SPECTRE_VULNERABLE ||
- spectre_bhb_state != SPECTRE_MITIGATED)
- return;
- if (!new_state)
- pr_err("WARNING: %s", EBPF_WARN);
- }
- #endif
|