mirror of
https://github.com/torvalds/linux.git
synced 2024-11-29 07:31:29 +00:00
7187bb7d0b
Cortex-X4 and Neoverse-V3 suffer from errata whereby an MSR to the SSBS special-purpose register does not affect subsequent speculative instructions, permitting speculative store bypassing for a window of time. This is described in their Software Developer Errata Notice (SDEN) documents: * Cortex-X4 SDEN v8.0, erratum 3194386: https://developer.arm.com/documentation/SDEN-2432808/0800/ * Neoverse-V3 SDEN v6.0, erratum 3312417: https://developer.arm.com/documentation/SDEN-2891958/0600/ To workaround these errata, it is necessary to place a speculation barrier (SB) after MSR to the SSBS special-purpose register. This patch adds the requisite SB after writes to SSBS within the kernel, and hides the presence of SSBS from EL0 such that userspace software which cares about SSBS will manipulate this via prctl(PR_GET_SPECULATION_CTRL, ...). Signed-off-by: Mark Rutland <mark.rutland@arm.com> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: James Morse <james.morse@arm.com> Cc: Will Deacon <will@kernel.org> Link: https://lore.kernel.org/r/20240508081400.235362-5-mark.rutland@arm.com Signed-off-by: Will Deacon <will@kernel.org>
769 lines
21 KiB
C
769 lines
21 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Contains CPU specific errata definitions
|
|
*
|
|
* Copyright (C) 2014 ARM Ltd.
|
|
*/
|
|
|
|
#include <linux/arm-smccc.h>
|
|
#include <linux/types.h>
|
|
#include <linux/cpu.h>
|
|
#include <asm/cpu.h>
|
|
#include <asm/cputype.h>
|
|
#include <asm/cpufeature.h>
|
|
#include <asm/kvm_asm.h>
|
|
#include <asm/smp_plat.h>
|
|
|
|
static bool __maybe_unused
|
|
is_affected_midr_range(const struct arm64_cpu_capabilities *entry, int scope)
|
|
{
|
|
const struct arm64_midr_revidr *fix;
|
|
u32 midr = read_cpuid_id(), revidr;
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
if (!is_midr_in_range(midr, &entry->midr_range))
|
|
return false;
|
|
|
|
midr &= MIDR_REVISION_MASK | MIDR_VARIANT_MASK;
|
|
revidr = read_cpuid(REVIDR_EL1);
|
|
for (fix = entry->fixed_revs; fix && fix->revidr_mask; fix++)
|
|
if (midr == fix->midr_rv && (revidr & fix->revidr_mask))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool __maybe_unused
|
|
is_affected_midr_range_list(const struct arm64_cpu_capabilities *entry,
|
|
int scope)
|
|
{
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
return is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list);
|
|
}
|
|
|
|
static bool __maybe_unused
|
|
is_kryo_midr(const struct arm64_cpu_capabilities *entry, int scope)
|
|
{
|
|
u32 model;
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
model = read_cpuid_id();
|
|
model &= MIDR_IMPLEMENTOR_MASK | (0xf00 << MIDR_PARTNUM_SHIFT) |
|
|
MIDR_ARCHITECTURE_MASK;
|
|
|
|
return model == entry->midr_range.model;
|
|
}
|
|
|
|
static bool
|
|
has_mismatched_cache_type(const struct arm64_cpu_capabilities *entry,
|
|
int scope)
|
|
{
|
|
u64 mask = arm64_ftr_reg_ctrel0.strict_mask;
|
|
u64 sys = arm64_ftr_reg_ctrel0.sys_val & mask;
|
|
u64 ctr_raw, ctr_real;
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
/*
|
|
* We want to make sure that all the CPUs in the system expose
|
|
* a consistent CTR_EL0 to make sure that applications behaves
|
|
* correctly with migration.
|
|
*
|
|
* If a CPU has CTR_EL0.IDC but does not advertise it via CTR_EL0 :
|
|
*
|
|
* 1) It is safe if the system doesn't support IDC, as CPU anyway
|
|
* reports IDC = 0, consistent with the rest.
|
|
*
|
|
* 2) If the system has IDC, it is still safe as we trap CTR_EL0
|
|
* access on this CPU via the ARM64_HAS_CACHE_IDC capability.
|
|
*
|
|
* So, we need to make sure either the raw CTR_EL0 or the effective
|
|
* CTR_EL0 matches the system's copy to allow a secondary CPU to boot.
|
|
*/
|
|
ctr_raw = read_cpuid_cachetype() & mask;
|
|
ctr_real = read_cpuid_effective_cachetype() & mask;
|
|
|
|
return (ctr_real != sys) && (ctr_raw != sys);
|
|
}
|
|
|
|
static void
|
|
cpu_enable_trap_ctr_access(const struct arm64_cpu_capabilities *cap)
|
|
{
|
|
u64 mask = arm64_ftr_reg_ctrel0.strict_mask;
|
|
bool enable_uct_trap = false;
|
|
|
|
/* Trap CTR_EL0 access on this CPU, only if it has a mismatch */
|
|
if ((read_cpuid_cachetype() & mask) !=
|
|
(arm64_ftr_reg_ctrel0.sys_val & mask))
|
|
enable_uct_trap = true;
|
|
|
|
/* ... or if the system is affected by an erratum */
|
|
if (cap->capability == ARM64_WORKAROUND_1542419)
|
|
enable_uct_trap = true;
|
|
|
|
if (enable_uct_trap)
|
|
sysreg_clear_set(sctlr_el1, SCTLR_EL1_UCT, 0);
|
|
}
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1463225
|
|
static bool
|
|
has_cortex_a76_erratum_1463225(const struct arm64_cpu_capabilities *entry,
|
|
int scope)
|
|
{
|
|
return is_affected_midr_range_list(entry, scope) && is_kernel_in_hyp_mode();
|
|
}
|
|
#endif
|
|
|
|
static void __maybe_unused
|
|
cpu_enable_cache_maint_trap(const struct arm64_cpu_capabilities *__unused)
|
|
{
|
|
sysreg_clear_set(sctlr_el1, SCTLR_EL1_UCI, 0);
|
|
}
|
|
|
|
#define CAP_MIDR_RANGE(model, v_min, r_min, v_max, r_max) \
|
|
.matches = is_affected_midr_range, \
|
|
.midr_range = MIDR_RANGE(model, v_min, r_min, v_max, r_max)
|
|
|
|
#define CAP_MIDR_ALL_VERSIONS(model) \
|
|
.matches = is_affected_midr_range, \
|
|
.midr_range = MIDR_ALL_VERSIONS(model)
|
|
|
|
#define MIDR_FIXED(rev, revidr_mask) \
|
|
.fixed_revs = (struct arm64_midr_revidr[]){{ (rev), (revidr_mask) }, {}}
|
|
|
|
#define ERRATA_MIDR_RANGE(model, v_min, r_min, v_max, r_max) \
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, \
|
|
CAP_MIDR_RANGE(model, v_min, r_min, v_max, r_max)
|
|
|
|
#define CAP_MIDR_RANGE_LIST(list) \
|
|
.matches = is_affected_midr_range_list, \
|
|
.midr_range_list = list
|
|
|
|
/* Errata affecting a range of revisions of given model variant */
|
|
#define ERRATA_MIDR_REV_RANGE(m, var, r_min, r_max) \
|
|
ERRATA_MIDR_RANGE(m, var, r_min, var, r_max)
|
|
|
|
/* Errata affecting a single variant/revision of a model */
|
|
#define ERRATA_MIDR_REV(model, var, rev) \
|
|
ERRATA_MIDR_RANGE(model, var, rev, var, rev)
|
|
|
|
/* Errata affecting all variants/revisions of a given a model */
|
|
#define ERRATA_MIDR_ALL_VERSIONS(model) \
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, \
|
|
CAP_MIDR_ALL_VERSIONS(model)
|
|
|
|
/* Errata affecting a list of midr ranges, with same work around */
|
|
#define ERRATA_MIDR_RANGE_LIST(midr_list) \
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, \
|
|
CAP_MIDR_RANGE_LIST(midr_list)
|
|
|
|
static const __maybe_unused struct midr_range tx2_family_cpus[] = {
|
|
MIDR_ALL_VERSIONS(MIDR_BRCM_VULCAN),
|
|
MIDR_ALL_VERSIONS(MIDR_CAVIUM_THUNDERX2),
|
|
{},
|
|
};
|
|
|
|
static bool __maybe_unused
|
|
needs_tx2_tvm_workaround(const struct arm64_cpu_capabilities *entry,
|
|
int scope)
|
|
{
|
|
int i;
|
|
|
|
if (!is_affected_midr_range_list(entry, scope) ||
|
|
!is_hyp_mode_available())
|
|
return false;
|
|
|
|
for_each_possible_cpu(i) {
|
|
if (MPIDR_AFFINITY_LEVEL(cpu_logical_map(i), 0) != 0)
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
static bool __maybe_unused
|
|
has_neoverse_n1_erratum_1542419(const struct arm64_cpu_capabilities *entry,
|
|
int scope)
|
|
{
|
|
u32 midr = read_cpuid_id();
|
|
bool has_dic = read_cpuid_cachetype() & BIT(CTR_EL0_DIC_SHIFT);
|
|
const struct midr_range range = MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N1);
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
return is_midr_in_range(midr, &range) && has_dic;
|
|
}
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_REPEAT_TLBI
|
|
static const struct arm64_cpu_capabilities arm64_repeat_tlbi_list[] = {
|
|
#ifdef CONFIG_QCOM_FALKOR_ERRATUM_1009
|
|
{
|
|
ERRATA_MIDR_REV(MIDR_QCOM_FALKOR_V1, 0, 0)
|
|
},
|
|
{
|
|
.midr_range.model = MIDR_QCOM_KRYO,
|
|
.matches = is_kryo_midr,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1286807
|
|
{
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 3, 0),
|
|
},
|
|
{
|
|
/* Kryo4xx Gold (rcpe to rfpe) => (r0p0 to r3p0) */
|
|
ERRATA_MIDR_RANGE(MIDR_QCOM_KRYO_4XX_GOLD, 0xc, 0xe, 0xf, 0xe),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2441007
|
|
{
|
|
ERRATA_MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2441009
|
|
{
|
|
/* Cortex-A510 r0p0 -> r1p1. Fixed in r1p2 */
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A510, 0, 0, 1, 1),
|
|
},
|
|
#endif
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_23154
|
|
static const struct midr_range cavium_erratum_23154_cpus[] = {
|
|
MIDR_ALL_VERSIONS(MIDR_THUNDERX),
|
|
MIDR_ALL_VERSIONS(MIDR_THUNDERX_81XX),
|
|
MIDR_ALL_VERSIONS(MIDR_THUNDERX_83XX),
|
|
MIDR_ALL_VERSIONS(MIDR_OCTX2_98XX),
|
|
MIDR_ALL_VERSIONS(MIDR_OCTX2_96XX),
|
|
MIDR_ALL_VERSIONS(MIDR_OCTX2_95XX),
|
|
MIDR_ALL_VERSIONS(MIDR_OCTX2_95XXN),
|
|
MIDR_ALL_VERSIONS(MIDR_OCTX2_95XXMM),
|
|
MIDR_ALL_VERSIONS(MIDR_OCTX2_95XXO),
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_27456
|
|
const struct midr_range cavium_erratum_27456_cpus[] = {
|
|
/* Cavium ThunderX, T88 pass 1.x - 2.1 */
|
|
MIDR_RANGE(MIDR_THUNDERX, 0, 0, 1, 1),
|
|
/* Cavium ThunderX, T81 pass 1.0 */
|
|
MIDR_REV(MIDR_THUNDERX_81XX, 0, 0),
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_30115
|
|
static const struct midr_range cavium_erratum_30115_cpus[] = {
|
|
/* Cavium ThunderX, T88 pass 1.x - 2.2 */
|
|
MIDR_RANGE(MIDR_THUNDERX, 0, 0, 1, 2),
|
|
/* Cavium ThunderX, T81 pass 1.0 - 1.2 */
|
|
MIDR_REV_RANGE(MIDR_THUNDERX_81XX, 0, 0, 2),
|
|
/* Cavium ThunderX, T83 pass 1.0 */
|
|
MIDR_REV(MIDR_THUNDERX_83XX, 0, 0),
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_QCOM_FALKOR_ERRATUM_1003
|
|
static const struct arm64_cpu_capabilities qcom_erratum_1003_list[] = {
|
|
{
|
|
ERRATA_MIDR_REV(MIDR_QCOM_FALKOR_V1, 0, 0),
|
|
},
|
|
{
|
|
.midr_range.model = MIDR_QCOM_KRYO,
|
|
.matches = is_kryo_midr,
|
|
},
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_CLEAN_CACHE
|
|
static const struct midr_range workaround_clean_cache[] = {
|
|
#if defined(CONFIG_ARM64_ERRATUM_826319) || \
|
|
defined(CONFIG_ARM64_ERRATUM_827319) || \
|
|
defined(CONFIG_ARM64_ERRATUM_824069)
|
|
/* Cortex-A53 r0p[012]: ARM errata 826319, 827319, 824069 */
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 2),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_819472
|
|
/* Cortex-A53 r0p[01] : ARM errata 819472 */
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 1),
|
|
#endif
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1418040
|
|
/*
|
|
* - 1188873 affects r0p0 to r2p0
|
|
* - 1418040 affects r0p0 to r3p1
|
|
*/
|
|
static const struct midr_range erratum_1418040_list[] = {
|
|
/* Cortex-A76 r0p0 to r3p1 */
|
|
MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 3, 1),
|
|
/* Neoverse-N1 r0p0 to r3p1 */
|
|
MIDR_RANGE(MIDR_NEOVERSE_N1, 0, 0, 3, 1),
|
|
/* Kryo4xx Gold (rcpe to rfpf) => (r0p0 to r3p1) */
|
|
MIDR_RANGE(MIDR_QCOM_KRYO_4XX_GOLD, 0xc, 0xe, 0xf, 0xf),
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_845719
|
|
static const struct midr_range erratum_845719_list[] = {
|
|
/* Cortex-A53 r0p[01234] */
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 4),
|
|
/* Brahma-B53 r0p[0] */
|
|
MIDR_REV(MIDR_BRAHMA_B53, 0, 0),
|
|
/* Kryo2XX Silver rAp4 */
|
|
MIDR_REV(MIDR_QCOM_KRYO_2XX_SILVER, 0xa, 0x4),
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_843419
|
|
static const struct arm64_cpu_capabilities erratum_843419_list[] = {
|
|
{
|
|
/* Cortex-A53 r0p[01234] */
|
|
.matches = is_affected_midr_range,
|
|
ERRATA_MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 4),
|
|
MIDR_FIXED(0x4, BIT(8)),
|
|
},
|
|
{
|
|
/* Brahma-B53 r0p[0] */
|
|
.matches = is_affected_midr_range,
|
|
ERRATA_MIDR_REV(MIDR_BRAHMA_B53, 0, 0),
|
|
},
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_AT
|
|
static const struct midr_range erratum_speculative_at_list[] = {
|
|
#ifdef CONFIG_ARM64_ERRATUM_1165522
|
|
/* Cortex A76 r0p0 to r2p0 */
|
|
MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 2, 0),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1319367
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A57),
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1530923
|
|
/* Cortex A55 r0p0 to r2p0 */
|
|
MIDR_RANGE(MIDR_CORTEX_A55, 0, 0, 2, 0),
|
|
/* Kryo4xx Silver (rdpe => r1p0) */
|
|
MIDR_REV(MIDR_QCOM_KRYO_4XX_SILVER, 0xd, 0xe),
|
|
#endif
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1463225
|
|
static const struct midr_range erratum_1463225[] = {
|
|
/* Cortex-A76 r0p0 - r3p1 */
|
|
MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 3, 1),
|
|
/* Kryo4xx Gold (rcpe to rfpf) => (r0p0 to r3p1) */
|
|
MIDR_RANGE(MIDR_QCOM_KRYO_4XX_GOLD, 0xc, 0xe, 0xf, 0xf),
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_TRBE_OVERWRITE_FILL_MODE
|
|
static const struct midr_range trbe_overwrite_fill_mode_cpus[] = {
|
|
#ifdef CONFIG_ARM64_ERRATUM_2139208
|
|
MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N2),
|
|
MIDR_ALL_VERSIONS(MIDR_MICROSOFT_AZURE_COBALT_100),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2119858
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A710),
|
|
MIDR_RANGE(MIDR_CORTEX_X2, 0, 0, 2, 0),
|
|
#endif
|
|
{},
|
|
};
|
|
#endif /* CONFIG_ARM64_WORKAROUND_TRBE_OVERWRITE_FILL_MODE */
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_TSB_FLUSH_FAILURE
|
|
static const struct midr_range tsb_flush_fail_cpus[] = {
|
|
#ifdef CONFIG_ARM64_ERRATUM_2067961
|
|
MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N2),
|
|
MIDR_ALL_VERSIONS(MIDR_MICROSOFT_AZURE_COBALT_100),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2054223
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A710),
|
|
#endif
|
|
{},
|
|
};
|
|
#endif /* CONFIG_ARM64_WORKAROUND_TSB_FLUSH_FAILURE */
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_TRBE_WRITE_OUT_OF_RANGE
|
|
static struct midr_range trbe_write_out_of_range_cpus[] = {
|
|
#ifdef CONFIG_ARM64_ERRATUM_2253138
|
|
MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N2),
|
|
MIDR_ALL_VERSIONS(MIDR_MICROSOFT_AZURE_COBALT_100),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2224489
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A710),
|
|
MIDR_RANGE(MIDR_CORTEX_X2, 0, 0, 2, 0),
|
|
#endif
|
|
{},
|
|
};
|
|
#endif /* CONFIG_ARM64_WORKAROUND_TRBE_WRITE_OUT_OF_RANGE */
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1742098
|
|
static struct midr_range broken_aarch32_aes[] = {
|
|
MIDR_RANGE(MIDR_CORTEX_A57, 0, 1, 0xf, 0xf),
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
|
|
{},
|
|
};
|
|
#endif /* CONFIG_ARM64_WORKAROUND_TRBE_WRITE_OUT_OF_RANGE */
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_UNPRIV_LOAD
|
|
static const struct midr_range erratum_spec_unpriv_load_list[] = {
|
|
#ifdef CONFIG_ARM64_ERRATUM_3117295
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A510),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2966298
|
|
/* Cortex-A520 r0p0 to r0p1 */
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A520, 0, 0, 1),
|
|
#endif
|
|
{},
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_SSBS
|
|
static const struct midr_range erratum_spec_ssbs_list[] = {
|
|
#ifdef CONFIG_ARM64_ERRATUM_3194386
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_X4),
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_3312417
|
|
MIDR_ALL_VERSIONS(MIDR_NEOVERSE_V3),
|
|
#endif
|
|
{}
|
|
};
|
|
#endif
|
|
|
|
const struct arm64_cpu_capabilities arm64_errata[] = {
|
|
#ifdef CONFIG_ARM64_WORKAROUND_CLEAN_CACHE
|
|
{
|
|
.desc = "ARM errata 826319, 827319, 824069, or 819472",
|
|
.capability = ARM64_WORKAROUND_CLEAN_CACHE,
|
|
ERRATA_MIDR_RANGE_LIST(workaround_clean_cache),
|
|
.cpu_enable = cpu_enable_cache_maint_trap,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_832075
|
|
{
|
|
/* Cortex-A57 r0p0 - r1p2 */
|
|
.desc = "ARM erratum 832075",
|
|
.capability = ARM64_WORKAROUND_DEVICE_LOAD_ACQUIRE,
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A57,
|
|
0, 0,
|
|
1, 2),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_834220
|
|
{
|
|
/* Cortex-A57 r0p0 - r1p2 */
|
|
.desc = "ARM erratum 834220",
|
|
.capability = ARM64_WORKAROUND_834220,
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A57,
|
|
0, 0,
|
|
1, 2),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_843419
|
|
{
|
|
.desc = "ARM erratum 843419",
|
|
.capability = ARM64_WORKAROUND_843419,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = cpucap_multi_entry_cap_matches,
|
|
.match_list = erratum_843419_list,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_845719
|
|
{
|
|
.desc = "ARM erratum 845719",
|
|
.capability = ARM64_WORKAROUND_845719,
|
|
ERRATA_MIDR_RANGE_LIST(erratum_845719_list),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_23154
|
|
{
|
|
.desc = "Cavium errata 23154 and 38545",
|
|
.capability = ARM64_WORKAROUND_CAVIUM_23154,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
ERRATA_MIDR_RANGE_LIST(cavium_erratum_23154_cpus),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_27456
|
|
{
|
|
.desc = "Cavium erratum 27456",
|
|
.capability = ARM64_WORKAROUND_CAVIUM_27456,
|
|
ERRATA_MIDR_RANGE_LIST(cavium_erratum_27456_cpus),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_30115
|
|
{
|
|
.desc = "Cavium erratum 30115",
|
|
.capability = ARM64_WORKAROUND_CAVIUM_30115,
|
|
ERRATA_MIDR_RANGE_LIST(cavium_erratum_30115_cpus),
|
|
},
|
|
#endif
|
|
{
|
|
.desc = "Mismatched cache type (CTR_EL0)",
|
|
.capability = ARM64_MISMATCHED_CACHE_TYPE,
|
|
.matches = has_mismatched_cache_type,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.cpu_enable = cpu_enable_trap_ctr_access,
|
|
},
|
|
#ifdef CONFIG_QCOM_FALKOR_ERRATUM_1003
|
|
{
|
|
.desc = "Qualcomm Technologies Falkor/Kryo erratum 1003",
|
|
.capability = ARM64_WORKAROUND_QCOM_FALKOR_E1003,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = cpucap_multi_entry_cap_matches,
|
|
.match_list = qcom_erratum_1003_list,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_REPEAT_TLBI
|
|
{
|
|
.desc = "Qualcomm erratum 1009, or ARM erratum 1286807, 2441009",
|
|
.capability = ARM64_WORKAROUND_REPEAT_TLBI,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = cpucap_multi_entry_cap_matches,
|
|
.match_list = arm64_repeat_tlbi_list,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_858921
|
|
{
|
|
/* Cortex-A73 all versions */
|
|
.desc = "ARM erratum 858921",
|
|
.capability = ARM64_WORKAROUND_858921,
|
|
ERRATA_MIDR_ALL_VERSIONS(MIDR_CORTEX_A73),
|
|
},
|
|
#endif
|
|
{
|
|
.desc = "Spectre-v2",
|
|
.capability = ARM64_SPECTRE_V2,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = has_spectre_v2,
|
|
.cpu_enable = spectre_v2_enable_mitigation,
|
|
},
|
|
#ifdef CONFIG_RANDOMIZE_BASE
|
|
{
|
|
/* Must come after the Spectre-v2 entry */
|
|
.desc = "Spectre-v3a",
|
|
.capability = ARM64_SPECTRE_V3A,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = has_spectre_v3a,
|
|
.cpu_enable = spectre_v3a_enable_mitigation,
|
|
},
|
|
#endif
|
|
{
|
|
.desc = "Spectre-v4",
|
|
.capability = ARM64_SPECTRE_V4,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = has_spectre_v4,
|
|
.cpu_enable = spectre_v4_enable_mitigation,
|
|
},
|
|
{
|
|
.desc = "Spectre-BHB",
|
|
.capability = ARM64_SPECTRE_BHB,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = is_spectre_bhb_affected,
|
|
.cpu_enable = spectre_bhb_enable_mitigation,
|
|
},
|
|
#ifdef CONFIG_ARM64_ERRATUM_1418040
|
|
{
|
|
.desc = "ARM erratum 1418040",
|
|
.capability = ARM64_WORKAROUND_1418040,
|
|
ERRATA_MIDR_RANGE_LIST(erratum_1418040_list),
|
|
/*
|
|
* We need to allow affected CPUs to come in late, but
|
|
* also need the non-affected CPUs to be able to come
|
|
* in at any point in time. Wonderful.
|
|
*/
|
|
.type = ARM64_CPUCAP_WEAK_LOCAL_CPU_FEATURE,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_AT
|
|
{
|
|
.desc = "ARM errata 1165522, 1319367, or 1530923",
|
|
.capability = ARM64_WORKAROUND_SPECULATIVE_AT,
|
|
ERRATA_MIDR_RANGE_LIST(erratum_speculative_at_list),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1463225
|
|
{
|
|
.desc = "ARM erratum 1463225",
|
|
.capability = ARM64_WORKAROUND_1463225,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = has_cortex_a76_erratum_1463225,
|
|
.midr_range_list = erratum_1463225,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_CAVIUM_TX2_ERRATUM_219
|
|
{
|
|
.desc = "Cavium ThunderX2 erratum 219 (KVM guest sysreg trapping)",
|
|
.capability = ARM64_WORKAROUND_CAVIUM_TX2_219_TVM,
|
|
ERRATA_MIDR_RANGE_LIST(tx2_family_cpus),
|
|
.matches = needs_tx2_tvm_workaround,
|
|
},
|
|
{
|
|
.desc = "Cavium ThunderX2 erratum 219 (PRFM removal)",
|
|
.capability = ARM64_WORKAROUND_CAVIUM_TX2_219_PRFM,
|
|
ERRATA_MIDR_RANGE_LIST(tx2_family_cpus),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1542419
|
|
{
|
|
/* we depend on the firmware portion for correctness */
|
|
.desc = "ARM erratum 1542419 (kernel portion)",
|
|
.capability = ARM64_WORKAROUND_1542419,
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
.matches = has_neoverse_n1_erratum_1542419,
|
|
.cpu_enable = cpu_enable_trap_ctr_access,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1508412
|
|
{
|
|
/* we depend on the firmware portion for correctness */
|
|
.desc = "ARM erratum 1508412 (kernel portion)",
|
|
.capability = ARM64_WORKAROUND_1508412,
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A77,
|
|
0, 0,
|
|
1, 0),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_NVIDIA_CARMEL_CNP_ERRATUM
|
|
{
|
|
/* NVIDIA Carmel */
|
|
.desc = "NVIDIA Carmel CNP erratum",
|
|
.capability = ARM64_WORKAROUND_NVIDIA_CARMEL_CNP,
|
|
ERRATA_MIDR_ALL_VERSIONS(MIDR_NVIDIA_CARMEL),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_TRBE_OVERWRITE_FILL_MODE
|
|
{
|
|
/*
|
|
* The erratum work around is handled within the TRBE
|
|
* driver and can be applied per-cpu. So, we can allow
|
|
* a late CPU to come online with this erratum.
|
|
*/
|
|
.desc = "ARM erratum 2119858 or 2139208",
|
|
.capability = ARM64_WORKAROUND_TRBE_OVERWRITE_FILL_MODE,
|
|
.type = ARM64_CPUCAP_WEAK_LOCAL_CPU_FEATURE,
|
|
CAP_MIDR_RANGE_LIST(trbe_overwrite_fill_mode_cpus),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_TSB_FLUSH_FAILURE
|
|
{
|
|
.desc = "ARM erratum 2067961 or 2054223",
|
|
.capability = ARM64_WORKAROUND_TSB_FLUSH_FAILURE,
|
|
ERRATA_MIDR_RANGE_LIST(tsb_flush_fail_cpus),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_TRBE_WRITE_OUT_OF_RANGE
|
|
{
|
|
.desc = "ARM erratum 2253138 or 2224489",
|
|
.capability = ARM64_WORKAROUND_TRBE_WRITE_OUT_OF_RANGE,
|
|
.type = ARM64_CPUCAP_WEAK_LOCAL_CPU_FEATURE,
|
|
CAP_MIDR_RANGE_LIST(trbe_write_out_of_range_cpus),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2645198
|
|
{
|
|
.desc = "ARM erratum 2645198",
|
|
.capability = ARM64_WORKAROUND_2645198,
|
|
ERRATA_MIDR_ALL_VERSIONS(MIDR_CORTEX_A715)
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2077057
|
|
{
|
|
.desc = "ARM erratum 2077057",
|
|
.capability = ARM64_WORKAROUND_2077057,
|
|
ERRATA_MIDR_REV_RANGE(MIDR_CORTEX_A510, 0, 0, 2),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2064142
|
|
{
|
|
.desc = "ARM erratum 2064142",
|
|
.capability = ARM64_WORKAROUND_2064142,
|
|
|
|
/* Cortex-A510 r0p0 - r0p2 */
|
|
ERRATA_MIDR_REV_RANGE(MIDR_CORTEX_A510, 0, 0, 2)
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2457168
|
|
{
|
|
.desc = "ARM erratum 2457168",
|
|
.capability = ARM64_WORKAROUND_2457168,
|
|
.type = ARM64_CPUCAP_WEAK_LOCAL_CPU_FEATURE,
|
|
|
|
/* Cortex-A510 r0p0-r1p1 */
|
|
CAP_MIDR_RANGE(MIDR_CORTEX_A510, 0, 0, 1, 1)
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2038923
|
|
{
|
|
.desc = "ARM erratum 2038923",
|
|
.capability = ARM64_WORKAROUND_2038923,
|
|
|
|
/* Cortex-A510 r0p0 - r0p2 */
|
|
ERRATA_MIDR_REV_RANGE(MIDR_CORTEX_A510, 0, 0, 2)
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1902691
|
|
{
|
|
.desc = "ARM erratum 1902691",
|
|
.capability = ARM64_WORKAROUND_1902691,
|
|
|
|
/* Cortex-A510 r0p0 - r0p1 */
|
|
ERRATA_MIDR_REV_RANGE(MIDR_CORTEX_A510, 0, 0, 1)
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_1742098
|
|
{
|
|
.desc = "ARM erratum 1742098",
|
|
.capability = ARM64_WORKAROUND_1742098,
|
|
CAP_MIDR_RANGE_LIST(broken_aarch32_aes),
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_ERRATUM_2658417
|
|
{
|
|
.desc = "ARM erratum 2658417",
|
|
.capability = ARM64_WORKAROUND_2658417,
|
|
/* Cortex-A510 r0p0 - r1p1 */
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A510, 0, 0, 1, 1),
|
|
MIDR_FIXED(MIDR_CPU_VAR_REV(1,1), BIT(25)),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_SSBS
|
|
{
|
|
.desc = "ARM errata 3194386, 3312417",
|
|
.capability = ARM64_WORKAROUND_SPECULATIVE_SSBS,
|
|
ERRATA_MIDR_RANGE_LIST(erratum_spec_ssbs_list),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_UNPRIV_LOAD
|
|
{
|
|
.desc = "ARM errata 2966298, 3117295",
|
|
.capability = ARM64_WORKAROUND_SPECULATIVE_UNPRIV_LOAD,
|
|
/* Cortex-A520 r0p0 - r0p1 */
|
|
ERRATA_MIDR_RANGE_LIST(erratum_spec_unpriv_load_list),
|
|
},
|
|
#endif
|
|
#ifdef CONFIG_AMPERE_ERRATUM_AC03_CPU_38
|
|
{
|
|
.desc = "AmpereOne erratum AC03_CPU_38",
|
|
.capability = ARM64_WORKAROUND_AMPERE_AC03_CPU_38,
|
|
ERRATA_MIDR_ALL_VERSIONS(MIDR_AMPERE1),
|
|
},
|
|
#endif
|
|
{
|
|
}
|
|
};
|