2019-06-03 05:44:50 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2012-03-05 11:49:27 +00:00
|
|
|
/*
|
|
|
|
* Based on arch/arm/mm/fault.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 1995 Linus Torvalds
|
|
|
|
* Copyright (C) 1995-2004 Russell King
|
|
|
|
* Copyright (C) 2012 ARM Ltd.
|
|
|
|
*/
|
|
|
|
|
2019-01-29 18:48:50 +00:00
|
|
|
#include <linux/acpi.h>
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
#include <linux/bitfield.h>
|
2016-09-19 21:38:55 +00:00
|
|
|
#include <linux/extable.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
#include <linux/signal.h>
|
|
|
|
#include <linux/mm.h>
|
|
|
|
#include <linux/hardirq.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/kprobes.h>
|
|
|
|
#include <linux/uaccess.h>
|
|
|
|
#include <linux/page-flags.h>
|
2017-02-08 17:51:30 +00:00
|
|
|
#include <linux/sched/signal.h>
|
2017-02-08 17:51:35 +00:00
|
|
|
#include <linux/sched/debug.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
#include <linux/highmem.h>
|
|
|
|
#include <linux/perf_event.h>
|
2016-10-18 10:27:47 +00:00
|
|
|
#include <linux/preempt.h>
|
2017-06-08 17:25:27 +00:00
|
|
|
#include <linux/hugetlb.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2019-01-29 18:48:50 +00:00
|
|
|
#include <asm/acpi.h>
|
2016-10-18 10:27:47 +00:00
|
|
|
#include <asm/bug.h>
|
2017-06-26 13:27:36 +00:00
|
|
|
#include <asm/cmpxchg.h>
|
2015-07-22 18:05:54 +00:00
|
|
|
#include <asm/cpufeature.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
#include <asm/exception.h>
|
2018-08-28 15:51:15 +00:00
|
|
|
#include <asm/daifflags.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
#include <asm/debug-monitors.h>
|
2014-04-06 22:04:12 +00:00
|
|
|
#include <asm/esr.h>
|
2019-10-25 16:42:10 +00:00
|
|
|
#include <asm/kprobes.h>
|
2019-10-25 16:42:16 +00:00
|
|
|
#include <asm/processor.h>
|
2015-07-22 18:05:54 +00:00
|
|
|
#include <asm/sysreg.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
#include <asm/system_misc.h>
|
|
|
|
#include <asm/tlbflush.h>
|
2018-02-20 14:53:22 +00:00
|
|
|
#include <asm/traps.h>
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2017-04-04 05:51:01 +00:00
|
|
|
struct fault_info {
|
|
|
|
int (*fn)(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs);
|
|
|
|
int sig;
|
|
|
|
int code;
|
|
|
|
const char *name;
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct fault_info fault_info[];
|
2018-09-22 15:39:54 +00:00
|
|
|
static struct fault_info debug_fault_info[];
|
2017-04-04 05:51:01 +00:00
|
|
|
|
|
|
|
static inline const struct fault_info *esr_to_fault_info(unsigned int esr)
|
|
|
|
{
|
2018-09-22 15:39:52 +00:00
|
|
|
return fault_info + (esr & ESR_ELx_FSC);
|
2017-04-04 05:51:01 +00:00
|
|
|
}
|
2012-10-24 15:34:02 +00:00
|
|
|
|
2018-09-22 15:39:54 +00:00
|
|
|
static inline const struct fault_info *esr_to_debug_fault_info(unsigned int esr)
|
|
|
|
{
|
|
|
|
return debug_fault_info + DBG_ESR_EVT(esr);
|
|
|
|
}
|
|
|
|
|
2017-08-04 08:31:42 +00:00
|
|
|
static void data_abort_decode(unsigned int esr)
|
|
|
|
{
|
|
|
|
pr_alert("Data abort info:\n");
|
|
|
|
|
|
|
|
if (esr & ESR_ELx_ISV) {
|
|
|
|
pr_alert(" Access size = %u byte(s)\n",
|
|
|
|
1U << ((esr & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT));
|
|
|
|
pr_alert(" SSE = %lu, SRT = %lu\n",
|
|
|
|
(esr & ESR_ELx_SSE) >> ESR_ELx_SSE_SHIFT,
|
|
|
|
(esr & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT);
|
|
|
|
pr_alert(" SF = %lu, AR = %lu\n",
|
|
|
|
(esr & ESR_ELx_SF) >> ESR_ELx_SF_SHIFT,
|
|
|
|
(esr & ESR_ELx_AR) >> ESR_ELx_AR_SHIFT);
|
|
|
|
} else {
|
2017-10-02 11:42:00 +00:00
|
|
|
pr_alert(" ISV = 0, ISS = 0x%08lx\n", esr & ESR_ELx_ISS_MASK);
|
2017-08-04 08:31:42 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pr_alert(" CM = %lu, WnR = %lu\n",
|
|
|
|
(esr & ESR_ELx_CM) >> ESR_ELx_CM_SHIFT,
|
|
|
|
(esr & ESR_ELx_WNR) >> ESR_ELx_WNR_SHIFT);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void mem_abort_decode(unsigned int esr)
|
|
|
|
{
|
|
|
|
pr_alert("Mem abort info:\n");
|
|
|
|
|
2017-10-19 10:19:55 +00:00
|
|
|
pr_alert(" ESR = 0x%08x\n", esr);
|
arm64: mm: print hexadecimal EC value in mem_abort_decode()
This change prints the hexadecimal EC value in mem_abort_decode(),
which makes it easier to lookup the corresponding EC in
the ARM Architecture Reference Manual.
The commit 1f9b8936f36f ("arm64: Decode information from ESR upon mem
faults") prints useful information when memory abort occurs. It would
be easier to lookup "0x25" instead of "DABT" in the document. Then we
can check the corresponding ISS.
For example:
Current info Document
EC Exception class
"CP15 MCR/MRC" 0x3 "MCR or MRC access to CP15a..."
"ASIMD" 0x7 "Access to SIMD or floating-point..."
"DABT (current EL)" 0x25 "Data Abort taken without..."
...
Before:
Unable to handle kernel paging request at virtual address 000000000000c000
Mem abort info:
ESR = 0x96000046
Exception class = DABT (current EL), IL = 32 bits
SET = 0, FnV = 0
EA = 0, S1PTW = 0
Data abort info:
ISV = 0, ISS = 0x00000046
CM = 0, WnR = 1
After:
Unable to handle kernel paging request at virtual address 000000000000c000
Mem abort info:
ESR = 0x96000046
EC = 0x25: DABT (current EL), IL = 32 bits
SET = 0, FnV = 0
EA = 0, S1PTW = 0
Data abort info:
ISV = 0, ISS = 0x00000046
CM = 0, WnR = 1
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: James Morse <james.morse@arm.com>
Acked-by: Mark Rutland <Mark.rutland@arm.com>
Signed-off-by: Miles Chen <miles.chen@mediatek.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-07 00:33:36 +00:00
|
|
|
pr_alert(" EC = 0x%02lx: %s, IL = %u bits\n",
|
|
|
|
ESR_ELx_EC(esr), esr_get_class_string(esr),
|
2017-08-04 08:31:42 +00:00
|
|
|
(esr & ESR_ELx_IL) ? 32 : 16);
|
|
|
|
pr_alert(" SET = %lu, FnV = %lu\n",
|
|
|
|
(esr & ESR_ELx_SET_MASK) >> ESR_ELx_SET_SHIFT,
|
|
|
|
(esr & ESR_ELx_FnV) >> ESR_ELx_FnV_SHIFT);
|
|
|
|
pr_alert(" EA = %lu, S1PTW = %lu\n",
|
|
|
|
(esr & ESR_ELx_EA) >> ESR_ELx_EA_SHIFT,
|
|
|
|
(esr & ESR_ELx_S1PTW) >> ESR_ELx_S1PTW_SHIFT);
|
|
|
|
|
|
|
|
if (esr_is_data_abort(esr))
|
|
|
|
data_abort_decode(esr);
|
|
|
|
}
|
|
|
|
|
arm64: mm: avoid virt_to_phys(init_mm.pgd)
If we take an unhandled fault in the kernel, we call show_pte() to dump
the {PGDP,PGD,PUD,PMD,PTE} values for the corresponding page table walk,
where the PGDP value is virt_to_phys(mm->pgd).
The boot-time and runtime kernel page tables, init_pg_dir and
swapper_pg_dir respectively, are kernel symbols. Thus, it is not valid
to call virt_to_phys() on either of these, though we'll do so if we take
a fault on a TTBR1 address.
When CONFIG_DEBUG_VIRTUAL is not selected, virt_to_phys() will silently
fix this up. However, when CONFIG_DEBUG_VIRTUAL is selected, this
results in splats as below. Depending on when these occur, they can
happen to suppress information needed to debug the original unhandled
fault, such as the backtrace:
| Unable to handle kernel paging request at virtual address ffff7fffec73cf0f
| Mem abort info:
| ESR = 0x96000004
| EC = 0x25: DABT (current EL), IL = 32 bits
| SET = 0, FnV = 0
| EA = 0, S1PTW = 0
| Data abort info:
| ISV = 0, ISS = 0x00000004
| CM = 0, WnR = 0
| ------------[ cut here ]------------
| virt_to_phys used for non-linear address: 00000000102c9dbe (swapper_pg_dir+0x0/0x1000)
| WARNING: CPU: 1 PID: 7558 at arch/arm64/mm/physaddr.c:15 __virt_to_phys+0xe0/0x170 arch/arm64/mm/physaddr.c:12
| Kernel panic - not syncing: panic_on_warn set ...
| SMP: stopping secondary CPUs
| Dumping ftrace buffer:
| (ftrace buffer empty)
| Kernel Offset: disabled
| CPU features: 0x0002,23000438
| Memory Limit: none
| Rebooting in 1 seconds..
We can avoid this by ensuring that we call __pa_symbol() for
init_mm.pgd, as this will always be a kernel symbol. As the dumped
{PGD,PUD,PMD,PTE} values are the raw values from the relevant entries we
don't need to handle these specially.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Cc: James Morse <james.morse@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-10-03 09:49:32 +00:00
|
|
|
static inline unsigned long mm_to_pgd_phys(struct mm_struct *mm)
|
|
|
|
{
|
|
|
|
/* Either init_pg_dir or swapper_pg_dir */
|
|
|
|
if (mm == &init_mm)
|
|
|
|
return __pa_symbol(mm->pgd);
|
|
|
|
|
|
|
|
return (unsigned long)virt_to_phys(mm->pgd);
|
|
|
|
}
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
/*
|
2017-06-09 15:35:52 +00:00
|
|
|
* Dump out the page tables associated with 'addr' in the currently active mm.
|
2012-03-05 11:49:27 +00:00
|
|
|
*/
|
2019-04-03 12:36:54 +00:00
|
|
|
static void show_pte(unsigned long addr)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2017-06-09 15:35:52 +00:00
|
|
|
struct mm_struct *mm;
|
2018-02-15 11:14:56 +00:00
|
|
|
pgd_t *pgdp;
|
|
|
|
pgd_t pgd;
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2018-12-28 08:30:27 +00:00
|
|
|
if (is_ttbr0_addr(addr)) {
|
2017-06-09 15:35:52 +00:00
|
|
|
/* TTBR0 */
|
|
|
|
mm = current->active_mm;
|
|
|
|
if (mm == &init_mm) {
|
|
|
|
pr_alert("[%016lx] user address but active_mm is swapper\n",
|
|
|
|
addr);
|
|
|
|
return;
|
|
|
|
}
|
2018-12-28 08:30:27 +00:00
|
|
|
} else if (is_ttbr1_addr(addr)) {
|
2017-06-09 15:35:52 +00:00
|
|
|
/* TTBR1 */
|
2012-03-05 11:49:27 +00:00
|
|
|
mm = &init_mm;
|
2017-06-09 15:35:52 +00:00
|
|
|
} else {
|
|
|
|
pr_alert("[%016lx] address between user and kernel address ranges\n",
|
|
|
|
addr);
|
|
|
|
return;
|
|
|
|
}
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2019-08-07 15:55:18 +00:00
|
|
|
pr_alert("%s pgtable: %luk pages, %llu-bit VAs, pgdp=%016lx\n",
|
2017-05-15 14:23:58 +00:00
|
|
|
mm == &init_mm ? "swapper" : "user", PAGE_SIZE / SZ_1K,
|
arm64: mm: avoid virt_to_phys(init_mm.pgd)
If we take an unhandled fault in the kernel, we call show_pte() to dump
the {PGDP,PGD,PUD,PMD,PTE} values for the corresponding page table walk,
where the PGDP value is virt_to_phys(mm->pgd).
The boot-time and runtime kernel page tables, init_pg_dir and
swapper_pg_dir respectively, are kernel symbols. Thus, it is not valid
to call virt_to_phys() on either of these, though we'll do so if we take
a fault on a TTBR1 address.
When CONFIG_DEBUG_VIRTUAL is not selected, virt_to_phys() will silently
fix this up. However, when CONFIG_DEBUG_VIRTUAL is selected, this
results in splats as below. Depending on when these occur, they can
happen to suppress information needed to debug the original unhandled
fault, such as the backtrace:
| Unable to handle kernel paging request at virtual address ffff7fffec73cf0f
| Mem abort info:
| ESR = 0x96000004
| EC = 0x25: DABT (current EL), IL = 32 bits
| SET = 0, FnV = 0
| EA = 0, S1PTW = 0
| Data abort info:
| ISV = 0, ISS = 0x00000004
| CM = 0, WnR = 0
| ------------[ cut here ]------------
| virt_to_phys used for non-linear address: 00000000102c9dbe (swapper_pg_dir+0x0/0x1000)
| WARNING: CPU: 1 PID: 7558 at arch/arm64/mm/physaddr.c:15 __virt_to_phys+0xe0/0x170 arch/arm64/mm/physaddr.c:12
| Kernel panic - not syncing: panic_on_warn set ...
| SMP: stopping secondary CPUs
| Dumping ftrace buffer:
| (ftrace buffer empty)
| Kernel Offset: disabled
| CPU features: 0x0002,23000438
| Memory Limit: none
| Rebooting in 1 seconds..
We can avoid this by ensuring that we call __pa_symbol() for
init_mm.pgd, as this will always be a kernel symbol. As the dumped
{PGD,PUD,PMD,PTE} values are the raw values from the relevant entries we
don't need to handle these specially.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Cc: James Morse <james.morse@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-10-03 09:49:32 +00:00
|
|
|
vabits_actual, mm_to_pgd_phys(mm));
|
2018-02-15 11:14:56 +00:00
|
|
|
pgdp = pgd_offset(mm, addr);
|
|
|
|
pgd = READ_ONCE(*pgdp);
|
|
|
|
pr_alert("[%016lx] pgd=%016llx", addr, pgd_val(pgd));
|
2012-03-05 11:49:27 +00:00
|
|
|
|
|
|
|
do {
|
2020-06-04 23:46:23 +00:00
|
|
|
p4d_t *p4dp, p4d;
|
2018-02-15 11:14:56 +00:00
|
|
|
pud_t *pudp, pud;
|
|
|
|
pmd_t *pmdp, pmd;
|
|
|
|
pte_t *ptep, pte;
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
if (pgd_none(pgd) || pgd_bad(pgd))
|
2012-03-05 11:49:27 +00:00
|
|
|
break;
|
|
|
|
|
2020-06-04 23:46:23 +00:00
|
|
|
p4dp = p4d_offset(pgdp, addr);
|
|
|
|
p4d = READ_ONCE(*p4dp);
|
|
|
|
pr_cont(", p4d=%016llx", p4d_val(p4d));
|
|
|
|
if (p4d_none(p4d) || p4d_bad(p4d))
|
|
|
|
break;
|
|
|
|
|
|
|
|
pudp = pud_offset(p4dp, addr);
|
2018-02-15 11:14:56 +00:00
|
|
|
pud = READ_ONCE(*pudp);
|
|
|
|
pr_cont(", pud=%016llx", pud_val(pud));
|
|
|
|
if (pud_none(pud) || pud_bad(pud))
|
2012-03-05 11:49:27 +00:00
|
|
|
break;
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
pmdp = pmd_offset(pudp, addr);
|
|
|
|
pmd = READ_ONCE(*pmdp);
|
|
|
|
pr_cont(", pmd=%016llx", pmd_val(pmd));
|
|
|
|
if (pmd_none(pmd) || pmd_bad(pmd))
|
2012-03-05 11:49:27 +00:00
|
|
|
break;
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
ptep = pte_offset_map(pmdp, addr);
|
|
|
|
pte = READ_ONCE(*ptep);
|
|
|
|
pr_cont(", pte=%016llx", pte_val(pte));
|
|
|
|
pte_unmap(ptep);
|
2012-03-05 11:49:27 +00:00
|
|
|
} while(0);
|
|
|
|
|
2017-01-03 14:27:26 +00:00
|
|
|
pr_cont("\n");
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 15:01:22 +00:00
|
|
|
/*
|
|
|
|
* This function sets the access flags (dirty, accessed), as well as write
|
|
|
|
* permission, and only to a more permissive setting.
|
|
|
|
*
|
|
|
|
* It needs to cope with hardware update of the accessed/dirty state by other
|
|
|
|
* agents in the system and can safely skip the __sync_icache_dcache() call as,
|
|
|
|
* like set_pte_at(), the PTE is never changed from no-exec to exec here.
|
|
|
|
*
|
|
|
|
* Returns whether or not the PTE actually changed.
|
|
|
|
*/
|
|
|
|
int ptep_set_access_flags(struct vm_area_struct *vma,
|
|
|
|
unsigned long address, pte_t *ptep,
|
|
|
|
pte_t entry, int dirty)
|
|
|
|
{
|
2017-06-26 13:27:36 +00:00
|
|
|
pteval_t old_pteval, pteval;
|
2018-02-15 11:14:56 +00:00
|
|
|
pte_t pte = READ_ONCE(*ptep);
|
2016-04-13 15:01:22 +00:00
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
if (pte_same(pte, entry))
|
2016-04-13 15:01:22 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* only preserve the access flags and write permission */
|
2017-07-04 18:04:18 +00:00
|
|
|
pte_val(entry) &= PTE_RDONLY | PTE_AF | PTE_WRITE | PTE_DIRTY;
|
2016-04-13 15:01:22 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Setting the flags must be done atomically to avoid racing with the
|
2017-07-25 13:53:03 +00:00
|
|
|
* hardware update of the access/dirty state. The PTE_RDONLY bit must
|
|
|
|
* be set to the most permissive (lowest value) of *ptep and entry
|
|
|
|
* (calculated as: a & b == ~(~a | ~b)).
|
2016-04-13 15:01:22 +00:00
|
|
|
*/
|
2017-07-25 13:53:03 +00:00
|
|
|
pte_val(entry) ^= PTE_RDONLY;
|
2018-02-15 11:14:56 +00:00
|
|
|
pteval = pte_val(pte);
|
2017-06-26 13:27:36 +00:00
|
|
|
do {
|
|
|
|
old_pteval = pteval;
|
|
|
|
pteval ^= PTE_RDONLY;
|
|
|
|
pteval |= pte_val(entry);
|
|
|
|
pteval ^= PTE_RDONLY;
|
|
|
|
pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval);
|
|
|
|
} while (pteval != old_pteval);
|
2016-04-13 15:01:22 +00:00
|
|
|
|
2020-09-30 12:20:40 +00:00
|
|
|
/* Invalidate a stale read-only entry */
|
|
|
|
if (dirty)
|
|
|
|
flush_tlb_page(vma, address);
|
2016-04-13 15:01:22 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2016-08-10 01:25:26 +00:00
|
|
|
static bool is_el1_instruction_abort(unsigned int esr)
|
|
|
|
{
|
|
|
|
return ESR_ELx_EC(esr) == ESR_ELx_EC_IABT_CUR;
|
|
|
|
}
|
|
|
|
|
2018-09-22 15:39:53 +00:00
|
|
|
static inline bool is_el1_permission_fault(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
2017-04-05 19:18:31 +00:00
|
|
|
{
|
|
|
|
unsigned int ec = ESR_ELx_EC(esr);
|
|
|
|
unsigned int fsc_type = esr & ESR_ELx_FSC_TYPE;
|
|
|
|
|
|
|
|
if (ec != ESR_ELx_EC_DABT_CUR && ec != ESR_ELx_EC_IABT_CUR)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (fsc_type == ESR_ELx_FSC_PERM)
|
|
|
|
return true;
|
|
|
|
|
2018-12-28 08:30:27 +00:00
|
|
|
if (is_ttbr0_addr(addr) && system_uses_ttbr0_pan())
|
2017-04-05 19:18:31 +00:00
|
|
|
return fsc_type == ESR_ELx_FSC_FAULT &&
|
|
|
|
(regs->pstate & PSR_PAN_BIT);
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
static bool __kprobes is_spurious_el1_translation_fault(unsigned long addr,
|
|
|
|
unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
u64 par, dfsc;
|
|
|
|
|
|
|
|
if (ESR_ELx_EC(esr) != ESR_ELx_EC_DABT_CUR ||
|
|
|
|
(esr & ESR_ELx_FSC_TYPE) != ESR_ELx_FSC_FAULT)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
local_irq_save(flags);
|
|
|
|
asm volatile("at s1e1r, %0" :: "r" (addr));
|
|
|
|
isb();
|
2020-10-28 18:28:39 +00:00
|
|
|
par = read_sysreg_par();
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
local_irq_restore(flags);
|
|
|
|
|
2019-10-16 11:03:04 +00:00
|
|
|
/*
|
|
|
|
* If we now have a valid translation, treat the translation fault as
|
|
|
|
* spurious.
|
|
|
|
*/
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
if (!(par & SYS_PAR_EL1_F))
|
2019-10-16 11:03:04 +00:00
|
|
|
return true;
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If we got a different type of fault from the AT instruction,
|
|
|
|
* treat the translation fault as spurious.
|
|
|
|
*/
|
2019-10-04 13:58:47 +00:00
|
|
|
dfsc = FIELD_GET(SYS_PAR_EL1_FST, par);
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
return (dfsc & ESR_ELx_FSC_TYPE) != ESR_ELx_FSC_FAULT;
|
|
|
|
}
|
|
|
|
|
2018-05-21 13:14:51 +00:00
|
|
|
static void die_kernel_fault(const char *msg, unsigned long addr,
|
|
|
|
unsigned int esr, struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
bust_spinlocks(1);
|
|
|
|
|
|
|
|
pr_alert("Unable to handle kernel %s at virtual address %016lx\n", msg,
|
|
|
|
addr);
|
|
|
|
|
|
|
|
mem_abort_decode(esr);
|
|
|
|
|
|
|
|
show_pte(addr);
|
|
|
|
die("Oops", regs, esr);
|
|
|
|
bust_spinlocks(0);
|
|
|
|
do_exit(SIGKILL);
|
|
|
|
}
|
|
|
|
|
2017-06-09 15:35:52 +00:00
|
|
|
static void __do_kernel_fault(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2017-04-05 19:18:31 +00:00
|
|
|
const char *msg;
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
/*
|
|
|
|
* Are we prepared to handle this kernel fault?
|
2016-08-10 01:25:26 +00:00
|
|
|
* We are almost certainly not prepared to handle instruction faults.
|
2012-03-05 11:49:27 +00:00
|
|
|
*/
|
2016-08-10 01:25:26 +00:00
|
|
|
if (!is_el1_instruction_abort(esr) && fixup_exception(regs))
|
2012-03-05 11:49:27 +00:00
|
|
|
return;
|
|
|
|
|
arm64: mm: Ignore spurious translation faults taken from the kernel
Thanks to address translation being performed out of order with respect to
loads and stores, it is possible for a CPU to take a translation fault when
accessing a page that was mapped by a different CPU.
For example, in the case that one CPU maps a page and then sets a flag to
tell another CPU:
CPU 0
-----
MOV X0, <valid pte>
STR X0, [Xptep] // Store new PTE to page table
DSB ISHST
ISB
MOV X1, #1
STR X1, [Xflag] // Set the flag
CPU 1
-----
loop: LDAR X0, [Xflag] // Poll flag with Acquire semantics
CBZ X0, loop
LDR X1, [X2] // Translates using the new PTE
then the final load on CPU 1 can raise a translation fault because the
translation can be performed speculatively before the read of the flag and
marked as "faulting" by the CPU. This isn't quite as bad as it sounds
since, in reality, code such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
*ptr = vmalloc(size); if (*ptr)
spin_unlock(&lock); foo = **ptr;
spin_unlock(&lock);
will not trigger the fault because there is an address dependency on CPU 1
which prevents the speculative translation. However, more exotic code where
the virtual address is known ahead of time, such as:
CPU 0 CPU 1
----- -----
spin_lock(&lock); spin_lock(&lock);
set_fixmap(0, paddr, prot); if (mapped)
mapped = true; foo = *fix_to_virt(0);
spin_unlock(&lock); spin_unlock(&lock);
could fault. This can be avoided by any of:
* Introducing broadcast TLB maintenance on the map path
* Adding a DSB;ISB sequence after checking a flag which indicates
that a virtual address is now mapped
* Handling the spurious fault
Given that we have never observed a problem due to this under Linux and
future revisions of the architecture are being tightened so that
translation table walks are effectively ordered in the same way as explicit
memory accesses, we no longer treat spurious kernel faults as fatal if an
AT instruction indicates that the access does not trigger a translation
fault.
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will@kernel.org>
2019-08-22 16:22:14 +00:00
|
|
|
if (WARN_RATELIMIT(is_spurious_el1_translation_fault(addr, esr, regs),
|
|
|
|
"Ignoring spurious kernel translation fault at virtual address %016lx\n", addr))
|
|
|
|
return;
|
|
|
|
|
2018-09-22 15:39:53 +00:00
|
|
|
if (is_el1_permission_fault(addr, esr, regs)) {
|
2017-04-05 19:18:31 +00:00
|
|
|
if (esr & ESR_ELx_WNR)
|
|
|
|
msg = "write to read-only memory";
|
2019-10-29 12:41:31 +00:00
|
|
|
else if (is_el1_instruction_abort(esr))
|
|
|
|
msg = "execute from non-executable memory";
|
2017-04-05 19:18:31 +00:00
|
|
|
else
|
|
|
|
msg = "read from unreadable memory";
|
|
|
|
} else if (addr < PAGE_SIZE) {
|
|
|
|
msg = "NULL pointer dereference";
|
|
|
|
} else {
|
|
|
|
msg = "paging request";
|
|
|
|
}
|
|
|
|
|
2018-05-21 13:14:51 +00:00
|
|
|
die_kernel_fault(msg, addr, esr, regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
2018-09-22 07:37:55 +00:00
|
|
|
static void set_thread_esr(unsigned long address, unsigned int esr)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2018-09-22 07:37:55 +00:00
|
|
|
current->thread.fault_address = address;
|
2018-05-22 16:11:20 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If the faulting address is in the kernel, we must sanitize the ESR.
|
|
|
|
* From userspace's point of view, kernel-only mappings don't exist
|
|
|
|
* at all, so we report them as level 0 translation faults.
|
|
|
|
* (This is not quite the way that "no mapping there at all" behaves:
|
|
|
|
* an alignment fault not caused by the memory type would take
|
|
|
|
* precedence over translation fault for a real access to empty
|
|
|
|
* space. Unfortunately we can't easily distinguish "alignment fault
|
|
|
|
* not caused by memory type" from "alignment fault caused by memory
|
|
|
|
* type", so we ignore this wrinkle and just return the translation
|
|
|
|
* fault.)
|
|
|
|
*/
|
2018-12-28 08:30:27 +00:00
|
|
|
if (!is_ttbr0_addr(current->thread.fault_address)) {
|
2018-05-22 16:11:20 +00:00
|
|
|
switch (ESR_ELx_EC(esr)) {
|
|
|
|
case ESR_ELx_EC_DABT_LOW:
|
|
|
|
/*
|
|
|
|
* These bits provide only information about the
|
|
|
|
* faulting instruction, which userspace knows already.
|
|
|
|
* We explicitly clear bits which are architecturally
|
|
|
|
* RES0 in case they are given meanings in future.
|
|
|
|
* We always report the ESR as if the fault was taken
|
|
|
|
* to EL1 and so ISV and the bits in ISS[23:14] are
|
|
|
|
* clear. (In fact it always will be a fault to EL1.)
|
|
|
|
*/
|
|
|
|
esr &= ESR_ELx_EC_MASK | ESR_ELx_IL |
|
|
|
|
ESR_ELx_CM | ESR_ELx_WNR;
|
|
|
|
esr |= ESR_ELx_FSC_FAULT;
|
|
|
|
break;
|
|
|
|
case ESR_ELx_EC_IABT_LOW:
|
|
|
|
/*
|
|
|
|
* Claim a level 0 translation fault.
|
|
|
|
* All other bits are architecturally RES0 for faults
|
|
|
|
* reported with that DFSC value, so we clear them.
|
|
|
|
*/
|
|
|
|
esr &= ESR_ELx_EC_MASK | ESR_ELx_IL;
|
|
|
|
esr |= ESR_ELx_FSC_FAULT;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
/*
|
|
|
|
* This should never happen (entry.S only brings us
|
|
|
|
* into this code for insn and data aborts from a lower
|
|
|
|
* exception level). Fail safe by not providing an ESR
|
|
|
|
* context record at all.
|
|
|
|
*/
|
|
|
|
WARN(1, "ESR 0x%x is not DABT or IABT from EL0\n", esr);
|
|
|
|
esr = 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-02-20 14:53:22 +00:00
|
|
|
current->thread.fault_code = esr;
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
2013-09-16 14:18:28 +00:00
|
|
|
static void do_bad_area(unsigned long addr, unsigned int esr, struct pt_regs *regs)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
* If we are in kernel mode at this point, we have no context to
|
|
|
|
* handle this fault with.
|
|
|
|
*/
|
2017-04-04 05:51:01 +00:00
|
|
|
if (user_mode(regs)) {
|
2018-02-20 14:53:22 +00:00
|
|
|
const struct fault_info *inf = esr_to_fault_info(esr);
|
2018-04-17 20:26:37 +00:00
|
|
|
|
2018-09-22 08:05:41 +00:00
|
|
|
set_thread_esr(addr, esr);
|
2018-09-22 08:26:57 +00:00
|
|
|
arm64_force_sig_fault(inf->sig, inf->code, (void __user *)addr,
|
|
|
|
inf->name);
|
2018-02-20 14:53:22 +00:00
|
|
|
} else {
|
2017-06-09 15:35:52 +00:00
|
|
|
__do_kernel_fault(addr, esr, regs);
|
2018-02-20 14:53:22 +00:00
|
|
|
}
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#define VM_FAULT_BADMAP 0x010000
|
|
|
|
#define VM_FAULT_BADACCESS 0x020000
|
|
|
|
|
2018-08-17 22:44:47 +00:00
|
|
|
static vm_fault_t __do_page_fault(struct mm_struct *mm, unsigned long addr,
|
2020-08-12 01:37:57 +00:00
|
|
|
unsigned int mm_flags, unsigned long vm_flags,
|
|
|
|
struct pt_regs *regs)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2019-06-07 09:13:06 +00:00
|
|
|
struct vm_area_struct *vma = find_vma(mm, addr);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
|
|
|
if (unlikely(!vma))
|
2019-06-07 09:13:06 +00:00
|
|
|
return VM_FAULT_BADMAP;
|
2012-03-05 11:49:27 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Ok, we have a good vm_area for this memory access, so we can handle
|
|
|
|
* it.
|
|
|
|
*/
|
2019-06-07 09:13:06 +00:00
|
|
|
if (unlikely(vma->vm_start > addr)) {
|
|
|
|
if (!(vma->vm_flags & VM_GROWSDOWN))
|
|
|
|
return VM_FAULT_BADMAP;
|
|
|
|
if (expand_stack(vma, addr))
|
|
|
|
return VM_FAULT_BADMAP;
|
|
|
|
}
|
|
|
|
|
2013-07-19 14:37:12 +00:00
|
|
|
/*
|
|
|
|
* Check that the permissions on the VMA allow for the fault which
|
2016-08-11 17:44:50 +00:00
|
|
|
* occurred.
|
2013-07-19 14:37:12 +00:00
|
|
|
*/
|
2019-06-07 09:13:06 +00:00
|
|
|
if (!(vma->vm_flags & vm_flags))
|
|
|
|
return VM_FAULT_BADACCESS;
|
2020-08-12 01:37:57 +00:00
|
|
|
return handle_mm_fault(vma, addr & PAGE_MASK, mm_flags, regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
arm64: kill ESR_LNX_EXEC
Currently we treat ESR_EL1 bit 24 as software-defined for distinguishing
instruction aborts from data aborts, but this bit is architecturally
RES0 for instruction aborts, and could be allocated for an arbitrary
purpose in future. Additionally, we hard-code the value in entry.S
without the mnemonic, making the code difficult to understand.
Instead, remove ESR_LNX_EXEC, and distinguish aborts based on the esr,
which we already pass to the sole use of ESR_LNX_EXEC. A new helper,
is_el0_instruction_abort() is added to make the logic clear. Any
instruction aborts taken from EL1 will already have been handled by
bad_mode, so we need not handle that case in the helper.
For consistency, the existing permission_fault helper is renamed to
is_permission_fault, and the return type is changed to bool. There
should be no functional changes as the return value was a boolean
expression, and the result is only used in another boolean expression.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Cc: Dave P Martin <dave.martin@arm.com>
Cc: Huang Shijie <shijie.huang@arm.com>
Cc: James Morse <james.morse@arm.com>
Cc: Marc Zyngier <marc.zyngier@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-05-31 11:33:03 +00:00
|
|
|
static bool is_el0_instruction_abort(unsigned int esr)
|
|
|
|
{
|
|
|
|
return ESR_ELx_EC(esr) == ESR_ELx_EC_IABT_LOW;
|
|
|
|
}
|
|
|
|
|
2019-06-07 09:13:05 +00:00
|
|
|
/*
|
|
|
|
* Note: not valid for EL1 DC IVAC, but we never use that such that it
|
|
|
|
* should fault. EL0 cannot issue DC IVAC (undef).
|
|
|
|
*/
|
|
|
|
static bool is_write_abort(unsigned int esr)
|
|
|
|
{
|
|
|
|
return (esr & ESR_ELx_WNR) && !(esr & ESR_ELx_CM);
|
|
|
|
}
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
static int __kprobes do_page_fault(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
|
|
|
{
|
2018-09-22 08:16:42 +00:00
|
|
|
const struct fault_info *inf;
|
2019-06-03 06:41:23 +00:00
|
|
|
struct mm_struct *mm = current->mm;
|
2020-08-12 01:37:57 +00:00
|
|
|
vm_fault_t fault;
|
2020-04-10 21:33:09 +00:00
|
|
|
unsigned long vm_flags = VM_ACCESS_FLAGS;
|
2020-04-02 04:08:37 +00:00
|
|
|
unsigned int mm_flags = FAULT_FLAG_DEFAULT;
|
2013-07-19 14:37:12 +00:00
|
|
|
|
2019-07-16 23:28:00 +00:00
|
|
|
if (kprobe_page_fault(regs, esr))
|
arm64: Kprobes with single stepping support
Add support for basic kernel probes(kprobes) and jump probes
(jprobes) for ARM64.
Kprobes utilizes software breakpoint and single step debug
exceptions supported on ARM v8.
A software breakpoint is placed at the probe address to trap the
kernel execution into the kprobe handler.
ARM v8 supports enabling single stepping before the break exception
return (ERET), with next PC in exception return address (ELR_EL1). The
kprobe handler prepares an executable memory slot for out-of-line
execution with a copy of the original instruction being probed, and
enables single stepping. The PC is set to the out-of-line slot address
before the ERET. With this scheme, the instruction is executed with the
exact same register context except for the PC (and DAIF) registers.
Debug mask (PSTATE.D) is enabled only when single stepping a recursive
kprobe, e.g.: during kprobes reenter so that probed instruction can be
single stepped within the kprobe handler -exception- context.
The recursion depth of kprobe is always 2, i.e. upon probe re-entry,
any further re-entry is prevented by not calling handlers and the case
counted as a missed kprobe).
Single stepping from the x-o-l slot has a drawback for PC-relative accesses
like branching and symbolic literals access as the offset from the new PC
(slot address) may not be ensured to fit in the immediate value of
the opcode. Such instructions need simulation, so reject
probing them.
Instructions generating exceptions or cpu mode change are rejected
for probing.
Exclusive load/store instructions are rejected too. Additionally, the
code is checked to see if it is inside an exclusive load/store sequence
(code from Pratyush).
System instructions are mostly enabled for stepping, except MSR/MRS
accesses to "DAIF" flags in PSTATE, which are not safe for
probing.
This also changes arch/arm64/include/asm/ptrace.h to use
include/asm-generic/ptrace.h.
Thanks to Steve Capper and Pratyush Anand for several suggested
Changes.
Signed-off-by: Sandeepa Prabhu <sandeepa.s.prabhu@gmail.com>
Signed-off-by: David A. Long <dave.long@linaro.org>
Signed-off-by: Pratyush Anand <panand@redhat.com>
Acked-by: Masami Hiramatsu <mhiramat@kernel.org>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-07-08 16:35:48 +00:00
|
|
|
return 0;
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
/*
|
|
|
|
* If we're in an interrupt or have no user context, we must not take
|
|
|
|
* the fault.
|
|
|
|
*/
|
2015-05-11 15:52:11 +00:00
|
|
|
if (faulthandler_disabled() || !mm)
|
2012-03-05 11:49:27 +00:00
|
|
|
goto no_context;
|
|
|
|
|
2013-09-12 22:13:39 +00:00
|
|
|
if (user_mode(regs))
|
|
|
|
mm_flags |= FAULT_FLAG_USER;
|
|
|
|
|
arm64: kill ESR_LNX_EXEC
Currently we treat ESR_EL1 bit 24 as software-defined for distinguishing
instruction aborts from data aborts, but this bit is architecturally
RES0 for instruction aborts, and could be allocated for an arbitrary
purpose in future. Additionally, we hard-code the value in entry.S
without the mnemonic, making the code difficult to understand.
Instead, remove ESR_LNX_EXEC, and distinguish aborts based on the esr,
which we already pass to the sole use of ESR_LNX_EXEC. A new helper,
is_el0_instruction_abort() is added to make the logic clear. Any
instruction aborts taken from EL1 will already have been handled by
bad_mode, so we need not handle that case in the helper.
For consistency, the existing permission_fault helper is renamed to
is_permission_fault, and the return type is changed to bool. There
should be no functional changes as the return value was a boolean
expression, and the result is only used in another boolean expression.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Cc: Dave P Martin <dave.martin@arm.com>
Cc: Huang Shijie <shijie.huang@arm.com>
Cc: James Morse <james.morse@arm.com>
Cc: Marc Zyngier <marc.zyngier@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-05-31 11:33:03 +00:00
|
|
|
if (is_el0_instruction_abort(esr)) {
|
2013-09-12 22:13:39 +00:00
|
|
|
vm_flags = VM_EXEC;
|
2019-05-05 04:15:12 +00:00
|
|
|
mm_flags |= FAULT_FLAG_INSTRUCTION;
|
2019-06-07 09:13:05 +00:00
|
|
|
} else if (is_write_abort(esr)) {
|
2013-09-12 22:13:39 +00:00
|
|
|
vm_flags = VM_WRITE;
|
|
|
|
mm_flags |= FAULT_FLAG_WRITE;
|
|
|
|
}
|
|
|
|
|
2018-12-28 08:30:27 +00:00
|
|
|
if (is_ttbr0_addr(addr) && is_el1_permission_fault(addr, esr, regs)) {
|
2016-06-20 17:28:01 +00:00
|
|
|
/* regs->orig_addr_limit may be 0 if we entered from EL0 */
|
|
|
|
if (regs->orig_addr_limit == KERNEL_DS)
|
2018-05-21 13:14:51 +00:00
|
|
|
die_kernel_fault("access to user memory with fs=KERNEL_DS",
|
|
|
|
addr, esr, regs);
|
2016-02-05 14:58:50 +00:00
|
|
|
|
2016-08-10 01:25:26 +00:00
|
|
|
if (is_el1_instruction_abort(esr))
|
2018-05-21 13:14:51 +00:00
|
|
|
die_kernel_fault("execution of user memory",
|
|
|
|
addr, esr, regs);
|
2016-08-10 01:25:26 +00:00
|
|
|
|
2016-02-05 14:58:48 +00:00
|
|
|
if (!search_exception_tables(regs->pc))
|
2018-05-21 13:14:51 +00:00
|
|
|
die_kernel_fault("access to user memory outside uaccess routines",
|
|
|
|
addr, esr, regs);
|
2016-02-05 14:58:48 +00:00
|
|
|
}
|
2015-07-22 18:05:54 +00:00
|
|
|
|
2017-06-08 17:25:28 +00:00
|
|
|
perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, addr);
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
/*
|
|
|
|
* As per x86, we may deadlock here. However, since the kernel only
|
|
|
|
* validly references user space from well defined areas of the code,
|
|
|
|
* we can bug out early if this is from code which shouldn't.
|
|
|
|
*/
|
2020-06-09 04:33:25 +00:00
|
|
|
if (!mmap_read_trylock(mm)) {
|
2012-03-05 11:49:27 +00:00
|
|
|
if (!user_mode(regs) && !search_exception_tables(regs->pc))
|
|
|
|
goto no_context;
|
|
|
|
retry:
|
2020-06-09 04:33:25 +00:00
|
|
|
mmap_read_lock(mm);
|
2012-03-05 11:49:27 +00:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* The above down_read_trylock() might have succeeded in which
|
|
|
|
* case, we'll have missed the might_sleep() from down_read().
|
|
|
|
*/
|
|
|
|
might_sleep();
|
|
|
|
#ifdef CONFIG_DEBUG_VM
|
2019-06-03 06:41:22 +00:00
|
|
|
if (!user_mode(regs) && !search_exception_tables(regs->pc)) {
|
2020-06-09 04:33:25 +00:00
|
|
|
mmap_read_unlock(mm);
|
2012-03-05 11:49:27 +00:00
|
|
|
goto no_context;
|
2019-06-03 06:41:22 +00:00
|
|
|
}
|
2012-03-05 11:49:27 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2020-08-12 01:37:57 +00:00
|
|
|
fault = __do_page_fault(mm, addr, mm_flags, vm_flags, regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2020-04-02 04:08:18 +00:00
|
|
|
/* Quick path to respond to signals */
|
|
|
|
if (fault_signal_pending(fault, regs)) {
|
|
|
|
if (!user_mode(regs))
|
|
|
|
goto no_context;
|
|
|
|
return 0;
|
|
|
|
}
|
2017-06-08 17:25:28 +00:00
|
|
|
|
2020-04-02 04:08:18 +00:00
|
|
|
if (fault & VM_FAULT_RETRY) {
|
2017-06-08 17:25:28 +00:00
|
|
|
if (mm_flags & FAULT_FLAG_ALLOW_RETRY) {
|
|
|
|
mm_flags |= FAULT_FLAG_TRIED;
|
|
|
|
goto retry;
|
|
|
|
}
|
|
|
|
}
|
2020-06-09 04:33:25 +00:00
|
|
|
mmap_read_unlock(mm);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
|
|
|
/*
|
2017-06-08 17:25:28 +00:00
|
|
|
* Handle the "normal" (no error) case first.
|
2012-03-05 11:49:27 +00:00
|
|
|
*/
|
2017-06-08 17:25:28 +00:00
|
|
|
if (likely(!(fault & (VM_FAULT_ERROR | VM_FAULT_BADMAP |
|
2020-08-12 01:37:57 +00:00
|
|
|
VM_FAULT_BADACCESS))))
|
2012-03-05 11:49:27 +00:00
|
|
|
return 0;
|
|
|
|
|
2013-09-12 22:13:38 +00:00
|
|
|
/*
|
|
|
|
* If we are in kernel mode at this point, we have no context to
|
|
|
|
* handle this fault with.
|
|
|
|
*/
|
|
|
|
if (!user_mode(regs))
|
|
|
|
goto no_context;
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
if (fault & VM_FAULT_OOM) {
|
|
|
|
/*
|
|
|
|
* We ran out of memory, call the OOM killer, and return to
|
|
|
|
* userspace (which will retry the fault, or kill us if we got
|
|
|
|
* oom-killed).
|
|
|
|
*/
|
|
|
|
pagefault_out_of_memory();
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-09-22 08:16:42 +00:00
|
|
|
inf = esr_to_fault_info(esr);
|
2018-09-22 08:18:42 +00:00
|
|
|
set_thread_esr(addr, esr);
|
2012-03-05 11:49:27 +00:00
|
|
|
if (fault & VM_FAULT_SIGBUS) {
|
|
|
|
/*
|
|
|
|
* We had some memory, but were unable to successfully fix up
|
|
|
|
* this page fault.
|
|
|
|
*/
|
2018-09-22 08:26:57 +00:00
|
|
|
arm64_force_sig_fault(SIGBUS, BUS_ADRERR, (void __user *)addr,
|
|
|
|
inf->name);
|
2018-09-22 07:46:39 +00:00
|
|
|
} else if (fault & (VM_FAULT_HWPOISON_LARGE | VM_FAULT_HWPOISON)) {
|
|
|
|
unsigned int lsb;
|
|
|
|
|
|
|
|
lsb = PAGE_SHIFT;
|
|
|
|
if (fault & VM_FAULT_HWPOISON_LARGE)
|
|
|
|
lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault));
|
2018-02-20 14:53:22 +00:00
|
|
|
|
2018-09-22 08:37:15 +00:00
|
|
|
arm64_force_sig_mceerr(BUS_MCEERR_AR, (void __user *)addr, lsb,
|
|
|
|
inf->name);
|
2012-03-05 11:49:27 +00:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Something tried to access memory that isn't in our memory
|
|
|
|
* map.
|
|
|
|
*/
|
2018-09-22 08:26:57 +00:00
|
|
|
arm64_force_sig_fault(SIGSEGV,
|
|
|
|
fault == VM_FAULT_BADACCESS ? SEGV_ACCERR : SEGV_MAPERR,
|
|
|
|
(void __user *)addr,
|
|
|
|
inf->name);
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
no_context:
|
2017-06-09 15:35:52 +00:00
|
|
|
__do_kernel_fault(addr, esr, regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __kprobes do_translation_fault(unsigned long addr,
|
|
|
|
unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
|
|
|
{
|
2018-12-28 08:30:27 +00:00
|
|
|
if (is_ttbr0_addr(addr))
|
2012-03-05 11:49:27 +00:00
|
|
|
return do_page_fault(addr, esr, regs);
|
|
|
|
|
|
|
|
do_bad_area(addr, esr, regs);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-02-16 04:44:35 +00:00
|
|
|
static int do_alignment_fault(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
do_bad_area(addr, esr, regs);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-03-05 11:49:27 +00:00
|
|
|
static int do_bad(unsigned long addr, unsigned int esr, struct pt_regs *regs)
|
|
|
|
{
|
2017-09-22 10:01:26 +00:00
|
|
|
return 1; /* "fault" */
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
|
|
|
|
2017-06-21 18:17:08 +00:00
|
|
|
static int do_sea(unsigned long addr, unsigned int esr, struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
const struct fault_info *inf;
|
2018-09-21 15:24:40 +00:00
|
|
|
void __user *siaddr;
|
2017-06-21 18:17:08 +00:00
|
|
|
|
|
|
|
inf = esr_to_fault_info(esr);
|
|
|
|
|
2020-05-01 16:45:43 +00:00
|
|
|
if (user_mode(regs) && apei_claim_sea(regs) == 0) {
|
|
|
|
/*
|
|
|
|
* APEI claimed this as a firmware-first notification.
|
|
|
|
* Some processing deferred to task_work before ret_to_user().
|
|
|
|
*/
|
|
|
|
return 0;
|
|
|
|
}
|
2017-06-21 18:17:09 +00:00
|
|
|
|
2017-06-21 18:17:08 +00:00
|
|
|
if (esr & ESR_ELx_FnV)
|
2018-09-21 15:24:40 +00:00
|
|
|
siaddr = NULL;
|
2017-06-21 18:17:08 +00:00
|
|
|
else
|
2018-09-21 15:24:40 +00:00
|
|
|
siaddr = (void __user *)addr;
|
|
|
|
arm64_notify_die(inf->name, regs, inf->sig, inf->code, siaddr, esr);
|
2017-06-21 18:17:08 +00:00
|
|
|
|
2017-12-13 10:36:47 +00:00
|
|
|
return 0;
|
2017-06-21 18:17:08 +00:00
|
|
|
}
|
|
|
|
|
2019-09-16 10:51:17 +00:00
|
|
|
static int do_tag_check_fault(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
do_bad_area(addr, esr, regs);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-04-04 05:51:01 +00:00
|
|
|
static const struct fault_info fault_info[] = {
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "ttbr address size fault" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "level 1 address size fault" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "level 2 address size fault" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "level 3 address size fault" },
|
2014-11-21 14:22:22 +00:00
|
|
|
{ do_translation_fault, SIGSEGV, SEGV_MAPERR, "level 0 translation fault" },
|
2012-03-05 11:49:27 +00:00
|
|
|
{ do_translation_fault, SIGSEGV, SEGV_MAPERR, "level 1 translation fault" },
|
|
|
|
{ do_translation_fault, SIGSEGV, SEGV_MAPERR, "level 2 translation fault" },
|
2017-09-29 11:27:41 +00:00
|
|
|
{ do_translation_fault, SIGSEGV, SEGV_MAPERR, "level 3 translation fault" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 8" },
|
2013-04-10 12:48:00 +00:00
|
|
|
{ do_page_fault, SIGSEGV, SEGV_ACCERR, "level 1 access flag fault" },
|
|
|
|
{ do_page_fault, SIGSEGV, SEGV_ACCERR, "level 2 access flag fault" },
|
2012-03-05 11:49:27 +00:00
|
|
|
{ do_page_fault, SIGSEGV, SEGV_ACCERR, "level 3 access flag fault" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 12" },
|
2013-04-10 12:48:00 +00:00
|
|
|
{ do_page_fault, SIGSEGV, SEGV_ACCERR, "level 1 permission fault" },
|
|
|
|
{ do_page_fault, SIGSEGV, SEGV_ACCERR, "level 2 permission fault" },
|
2012-03-05 11:49:27 +00:00
|
|
|
{ do_page_fault, SIGSEGV, SEGV_ACCERR, "level 3 permission fault" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_sea, SIGBUS, BUS_OBJERR, "synchronous external abort" },
|
2019-09-16 10:51:17 +00:00
|
|
|
{ do_tag_check_fault, SIGSEGV, SEGV_MTESERR, "synchronous tag check fault" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 18" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 19" },
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 0 (translation table walk)" },
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 1 (translation table walk)" },
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 2 (translation table walk)" },
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 3 (translation table walk)" },
|
|
|
|
{ do_sea, SIGBUS, BUS_OBJERR, "synchronous parity or ECC error" }, // Reserved when RAS is implemented
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 25" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 26" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 27" },
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 0 synchronous parity error (translation table walk)" }, // Reserved when RAS is implemented
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 1 synchronous parity error (translation table walk)" }, // Reserved when RAS is implemented
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 2 synchronous parity error (translation table walk)" }, // Reserved when RAS is implemented
|
|
|
|
{ do_sea, SIGKILL, SI_KERNEL, "level 3 synchronous parity error (translation table walk)" }, // Reserved when RAS is implemented
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 32" },
|
2016-02-16 04:44:35 +00:00
|
|
|
{ do_alignment_fault, SIGBUS, BUS_ADRALN, "alignment fault" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 34" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 35" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 36" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 37" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 38" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 39" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 40" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 41" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 42" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 43" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 44" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 45" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 46" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 47" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "TLB conflict abort" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "Unsupported atomic hardware update fault" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 50" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 51" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "implementation fault (lockdown abort)" },
|
|
|
|
{ do_bad, SIGBUS, BUS_OBJERR, "implementation fault (unsupported exclusive)" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 54" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 55" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 56" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 57" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 58" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 59" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 60" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "section domain fault" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "page domain fault" },
|
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 63" },
|
2012-03-05 11:49:27 +00:00
|
|
|
};
|
|
|
|
|
2019-10-25 16:42:15 +00:00
|
|
|
void do_mem_abort(unsigned long addr, unsigned int esr, struct pt_regs *regs)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2017-04-04 05:51:01 +00:00
|
|
|
const struct fault_info *inf = esr_to_fault_info(esr);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
|
|
|
if (!inf->fn(addr, esr, regs))
|
|
|
|
return;
|
|
|
|
|
2018-02-20 14:41:02 +00:00
|
|
|
if (!user_mode(regs)) {
|
|
|
|
pr_alert("Unhandled fault at 0x%016lx\n", addr);
|
|
|
|
mem_abort_decode(esr);
|
2017-10-31 15:56:11 +00:00
|
|
|
show_pte(addr);
|
2018-02-20 14:41:02 +00:00
|
|
|
}
|
2017-10-19 10:19:55 +00:00
|
|
|
|
2018-09-21 15:24:40 +00:00
|
|
|
arm64_notify_die(inf->name, regs,
|
|
|
|
inf->sig, inf->code, (void __user *)addr, esr);
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
2019-10-25 16:42:10 +00:00
|
|
|
NOKPROBE_SYMBOL(do_mem_abort);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2019-10-25 16:42:15 +00:00
|
|
|
void do_el0_irq_bp_hardening(void)
|
2018-02-02 17:31:40 +00:00
|
|
|
{
|
|
|
|
/* PC has already been checked in entry.S */
|
|
|
|
arm64_apply_bp_hardening();
|
|
|
|
}
|
2019-10-25 16:42:10 +00:00
|
|
|
NOKPROBE_SYMBOL(do_el0_irq_bp_hardening);
|
2018-02-02 17:31:40 +00:00
|
|
|
|
2019-10-25 16:42:15 +00:00
|
|
|
void do_sp_pc_abort(unsigned long addr, unsigned int esr, struct pt_regs *regs)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2018-09-21 15:24:40 +00:00
|
|
|
arm64_notify_die("SP/PC alignment exception", regs,
|
|
|
|
SIGBUS, BUS_ADRALN, (void __user *)addr, esr);
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
2019-10-25 16:42:10 +00:00
|
|
|
NOKPROBE_SYMBOL(do_sp_pc_abort);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2015-07-24 15:37:48 +00:00
|
|
|
int __init early_brk64(unsigned long addr, unsigned int esr,
|
|
|
|
struct pt_regs *regs);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* __refdata because early_brk64 is __init, but the reference to it is
|
|
|
|
* clobbered at arch_initcall time.
|
|
|
|
* See traps.c and debug-monitors.c:debug_traps_init().
|
|
|
|
*/
|
|
|
|
static struct fault_info __refdata debug_fault_info[] = {
|
2012-03-05 11:49:27 +00:00
|
|
|
{ do_bad, SIGTRAP, TRAP_HWBKPT, "hardware breakpoint" },
|
|
|
|
{ do_bad, SIGTRAP, TRAP_HWBKPT, "hardware single-step" },
|
|
|
|
{ do_bad, SIGTRAP, TRAP_HWBKPT, "hardware watchpoint" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 3" },
|
2012-03-05 11:49:27 +00:00
|
|
|
{ do_bad, SIGTRAP, TRAP_BRKPT, "aarch32 BKPT" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "aarch32 vector catch" },
|
2015-07-24 15:37:48 +00:00
|
|
|
{ early_brk64, SIGTRAP, TRAP_BRKPT, "aarch64 BRK" },
|
arm64: signal: Ensure si_code is valid for all fault signals
Currently, as reported by Eric, an invalid si_code value 0 is
passed in many signals delivered to userspace in response to faults
and other kernel errors. Typically 0 is passed when the fault is
insufficiently diagnosable or when there does not appear to be any
sensible alternative value to choose.
This appears to violate POSIX, and is intuitively wrong for at
least two reasons arising from the fact that 0 == SI_USER:
1) si_code is a union selector, and SI_USER (and si_code <= 0 in
general) implies the existence of a different set of fields
(siginfo._kill) from that which exists for a fault signal
(siginfo._sigfault). However, the code raising the signal
typically writes only the _sigfault fields, and the _kill
fields make no sense in this case.
Thus when userspace sees si_code == 0 (SI_USER) it may
legitimately inspect fields in the inactive union member _kill
and obtain garbage as a result.
There appears to be software in the wild relying on this,
albeit generally only for printing diagnostic messages.
2) Software that wants to be robust against spurious signals may
discard signals where si_code == SI_USER (or <= 0), or may
filter such signals based on the si_uid and si_pid fields of
siginfo._sigkill. In the case of fault signals, this means
that important (and usually fatal) error conditions may be
silently ignored.
In practice, many of the faults for which arm64 passes si_code == 0
are undiagnosable conditions such as exceptions with syndrome
values in ESR_ELx to which the architecture does not yet assign any
meaning, or conditions indicative of a bug or error in the kernel
or system and thus that are unrecoverable and should never occur in
normal operation.
The approach taken in this patch is to translate all such
undiagnosable or "impossible" synchronous fault conditions to
SIGKILL, since these are at least probably localisable to a single
process. Some of these conditions should really result in a kernel
panic, but due to the lack of diagnostic information it is
difficult to be certain: this patch does not add any calls to
panic(), but this could change later if justified.
Although si_code will not reach userspace in the case of SIGKILL,
it is still desirable to pass a nonzero value so that the common
siginfo handling code can detect incorrect use of si_code == 0
without false positives. In this case the si_code dependent
siginfo fields will not be correctly initialised, but since they
are not passed to userspace I deem this not to matter.
A few faults can reasonably occur in realistic userspace scenarios,
and _should_ raise a regular, handleable (but perhaps not
ignorable/blockable) signal: for these, this patch attempts to
choose a suitable standard si_code value for the raised signal in
each case instead of 0.
arm64 was the only arch to define a BUS_FIXME code, so after this
patch nobody defines it. This patch therefore also removes the
relevant code from siginfo_layout().
Cc: James Morse <james.morse@arm.com>
Reported-by: Eric W. Biederman <ebiederm@xmission.com>
Signed-off-by: Dave Martin <Dave.Martin@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
2018-03-08 17:41:05 +00:00
|
|
|
{ do_bad, SIGKILL, SI_KERNEL, "unknown 7" },
|
2012-03-05 11:49:27 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
void __init hook_debug_fault_code(int nr,
|
|
|
|
int (*fn)(unsigned long, unsigned int, struct pt_regs *),
|
|
|
|
int sig, int code, const char *name)
|
|
|
|
{
|
|
|
|
BUG_ON(nr < 0 || nr >= ARRAY_SIZE(debug_fault_info));
|
|
|
|
|
|
|
|
debug_fault_info[nr].fn = fn;
|
|
|
|
debug_fault_info[nr].sig = sig;
|
|
|
|
debug_fault_info[nr].code = code;
|
|
|
|
debug_fault_info[nr].name = name;
|
|
|
|
}
|
|
|
|
|
2019-08-01 14:36:14 +00:00
|
|
|
/*
|
|
|
|
* In debug exception context, we explicitly disable preemption despite
|
|
|
|
* having interrupts disabled.
|
|
|
|
* This serves two purposes: it makes it much less likely that we would
|
|
|
|
* accidentally schedule in exception context and it will force a warning
|
|
|
|
* if we somehow manage to schedule by accident.
|
|
|
|
*/
|
|
|
|
static void debug_exception_enter(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Tell lockdep we disabled irqs in entry.S. Do nothing if they were
|
|
|
|
* already disabled to preserve the last enabled/disabled addresses.
|
|
|
|
*/
|
|
|
|
if (interrupts_enabled(regs))
|
|
|
|
trace_hardirqs_off();
|
|
|
|
|
|
|
|
if (user_mode(regs)) {
|
|
|
|
RCU_LOCKDEP_WARN(!rcu_is_watching(), "entry code didn't wake RCU");
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* We might have interrupted pretty much anything. In
|
|
|
|
* fact, if we're a debug exception, we can even interrupt
|
|
|
|
* NMI processing. We don't want this code makes in_nmi()
|
|
|
|
* to return true, but we need to notify RCU.
|
|
|
|
*/
|
|
|
|
rcu_nmi_enter();
|
|
|
|
}
|
|
|
|
|
|
|
|
preempt_disable();
|
|
|
|
|
|
|
|
/* This code is a bit fragile. Test it. */
|
|
|
|
RCU_LOCKDEP_WARN(!rcu_is_watching(), "exception_enter didn't work");
|
|
|
|
}
|
|
|
|
NOKPROBE_SYMBOL(debug_exception_enter);
|
|
|
|
|
|
|
|
static void debug_exception_exit(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
preempt_enable_no_resched();
|
|
|
|
|
|
|
|
if (!user_mode(regs))
|
|
|
|
rcu_nmi_exit();
|
|
|
|
|
|
|
|
if (interrupts_enabled(regs))
|
|
|
|
trace_hardirqs_on();
|
|
|
|
}
|
|
|
|
NOKPROBE_SYMBOL(debug_exception_exit);
|
|
|
|
|
2019-04-29 12:03:57 +00:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1463225
|
|
|
|
DECLARE_PER_CPU(int, __in_cortex_a76_erratum_1463225_wa);
|
|
|
|
|
2019-10-25 16:42:10 +00:00
|
|
|
static int cortex_a76_erratum_1463225_debug_handler(struct pt_regs *regs)
|
2019-04-29 12:03:57 +00:00
|
|
|
{
|
|
|
|
if (user_mode(regs))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!__this_cpu_read(__in_cortex_a76_erratum_1463225_wa))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We've taken a dummy step exception from the kernel to ensure
|
|
|
|
* that interrupts are re-enabled on the syscall path. Return back
|
|
|
|
* to cortex_a76_erratum_1463225_svc_handler() with debug exceptions
|
|
|
|
* masked so that we can safely restore the mdscr and get on with
|
|
|
|
* handling the syscall.
|
|
|
|
*/
|
|
|
|
regs->pstate |= PSR_D_BIT;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
#else
|
2019-10-25 16:42:10 +00:00
|
|
|
static int cortex_a76_erratum_1463225_debug_handler(struct pt_regs *regs)
|
2019-04-29 12:03:57 +00:00
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif /* CONFIG_ARM64_ERRATUM_1463225 */
|
2019-10-25 16:42:10 +00:00
|
|
|
NOKPROBE_SYMBOL(cortex_a76_erratum_1463225_debug_handler);
|
2019-04-29 12:03:57 +00:00
|
|
|
|
2019-10-25 16:42:15 +00:00
|
|
|
void do_debug_exception(unsigned long addr_if_watchpoint, unsigned int esr,
|
|
|
|
struct pt_regs *regs)
|
2012-03-05 11:49:27 +00:00
|
|
|
{
|
2018-09-22 15:39:54 +00:00
|
|
|
const struct fault_info *inf = esr_to_debug_fault_info(esr);
|
2019-03-01 13:28:00 +00:00
|
|
|
unsigned long pc = instruction_pointer(regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2019-04-29 12:03:57 +00:00
|
|
|
if (cortex_a76_erratum_1463225_debug_handler(regs))
|
|
|
|
return;
|
|
|
|
|
2019-08-01 14:36:14 +00:00
|
|
|
debug_exception_enter(regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2019-03-01 13:28:00 +00:00
|
|
|
if (user_mode(regs) && !is_ttbr0_addr(pc))
|
2018-02-02 17:31:39 +00:00
|
|
|
arm64_apply_bp_hardening();
|
|
|
|
|
2019-02-25 12:06:43 +00:00
|
|
|
if (inf->fn(addr_if_watchpoint, esr, regs)) {
|
2018-09-21 15:24:40 +00:00
|
|
|
arm64_notify_die(inf->name, regs,
|
2019-03-01 13:28:00 +00:00
|
|
|
inf->sig, inf->code, (void __user *)pc, esr);
|
2016-04-13 12:40:00 +00:00
|
|
|
}
|
2012-03-05 11:49:27 +00:00
|
|
|
|
2019-08-01 14:36:14 +00:00
|
|
|
debug_exception_exit(regs);
|
2012-03-05 11:49:27 +00:00
|
|
|
}
|
arm64: Kprobes with single stepping support
Add support for basic kernel probes(kprobes) and jump probes
(jprobes) for ARM64.
Kprobes utilizes software breakpoint and single step debug
exceptions supported on ARM v8.
A software breakpoint is placed at the probe address to trap the
kernel execution into the kprobe handler.
ARM v8 supports enabling single stepping before the break exception
return (ERET), with next PC in exception return address (ELR_EL1). The
kprobe handler prepares an executable memory slot for out-of-line
execution with a copy of the original instruction being probed, and
enables single stepping. The PC is set to the out-of-line slot address
before the ERET. With this scheme, the instruction is executed with the
exact same register context except for the PC (and DAIF) registers.
Debug mask (PSTATE.D) is enabled only when single stepping a recursive
kprobe, e.g.: during kprobes reenter so that probed instruction can be
single stepped within the kprobe handler -exception- context.
The recursion depth of kprobe is always 2, i.e. upon probe re-entry,
any further re-entry is prevented by not calling handlers and the case
counted as a missed kprobe).
Single stepping from the x-o-l slot has a drawback for PC-relative accesses
like branching and symbolic literals access as the offset from the new PC
(slot address) may not be ensured to fit in the immediate value of
the opcode. Such instructions need simulation, so reject
probing them.
Instructions generating exceptions or cpu mode change are rejected
for probing.
Exclusive load/store instructions are rejected too. Additionally, the
code is checked to see if it is inside an exclusive load/store sequence
(code from Pratyush).
System instructions are mostly enabled for stepping, except MSR/MRS
accesses to "DAIF" flags in PSTATE, which are not safe for
probing.
This also changes arch/arm64/include/asm/ptrace.h to use
include/asm-generic/ptrace.h.
Thanks to Steve Capper and Pratyush Anand for several suggested
Changes.
Signed-off-by: Sandeepa Prabhu <sandeepa.s.prabhu@gmail.com>
Signed-off-by: David A. Long <dave.long@linaro.org>
Signed-off-by: Pratyush Anand <panand@redhat.com>
Acked-by: Masami Hiramatsu <mhiramat@kernel.org>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-07-08 16:35:48 +00:00
|
|
|
NOKPROBE_SYMBOL(do_debug_exception);
|