mirror of
https://github.com/torvalds/linux.git
synced 2024-11-26 22:21:42 +00:00
048e2906d4
Implement arch_align_stack() to randomize the lower bits of the stack address. Signed-off-by: Yunhui Cui <cuiyunhui@bytedance.com> Link: https://lore.kernel.org/r/20240625030502.68988-1-cuiyunhui@bytedance.com Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
245 lines
6.4 KiB
C
245 lines
6.4 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2009 Sunplus Core Technology Co., Ltd.
|
|
* Chen Liqin <liqin.chen@sunplusct.com>
|
|
* Lennox Wu <lennox.wu@sunplusct.com>
|
|
* Copyright (C) 2012 Regents of the University of California
|
|
* Copyright (C) 2017 SiFive
|
|
*/
|
|
|
|
#include <linux/cpu.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/sched/debug.h>
|
|
#include <linux/sched/task_stack.h>
|
|
#include <linux/tick.h>
|
|
#include <linux/ptrace.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/personality.h>
|
|
|
|
#include <asm/unistd.h>
|
|
#include <asm/processor.h>
|
|
#include <asm/csr.h>
|
|
#include <asm/stacktrace.h>
|
|
#include <asm/string.h>
|
|
#include <asm/switch_to.h>
|
|
#include <asm/thread_info.h>
|
|
#include <asm/cpuidle.h>
|
|
#include <asm/vector.h>
|
|
#include <asm/cpufeature.h>
|
|
#include <asm/exec.h>
|
|
|
|
#if defined(CONFIG_STACKPROTECTOR) && !defined(CONFIG_STACKPROTECTOR_PER_TASK)
|
|
#include <linux/stackprotector.h>
|
|
unsigned long __stack_chk_guard __read_mostly;
|
|
EXPORT_SYMBOL(__stack_chk_guard);
|
|
#endif
|
|
|
|
extern asmlinkage void ret_from_fork(void);
|
|
|
|
void noinstr arch_cpu_idle(void)
|
|
{
|
|
cpu_do_idle();
|
|
}
|
|
|
|
int set_unalign_ctl(struct task_struct *tsk, unsigned int val)
|
|
{
|
|
if (!unaligned_ctl_available())
|
|
return -EINVAL;
|
|
|
|
tsk->thread.align_ctl = val;
|
|
return 0;
|
|
}
|
|
|
|
int get_unalign_ctl(struct task_struct *tsk, unsigned long adr)
|
|
{
|
|
if (!unaligned_ctl_available())
|
|
return -EINVAL;
|
|
|
|
return put_user(tsk->thread.align_ctl, (unsigned long __user *)adr);
|
|
}
|
|
|
|
void __show_regs(struct pt_regs *regs)
|
|
{
|
|
show_regs_print_info(KERN_DEFAULT);
|
|
|
|
if (!user_mode(regs)) {
|
|
pr_cont("epc : %pS\n", (void *)regs->epc);
|
|
pr_cont(" ra : %pS\n", (void *)regs->ra);
|
|
}
|
|
|
|
pr_cont("epc : " REG_FMT " ra : " REG_FMT " sp : " REG_FMT "\n",
|
|
regs->epc, regs->ra, regs->sp);
|
|
pr_cont(" gp : " REG_FMT " tp : " REG_FMT " t0 : " REG_FMT "\n",
|
|
regs->gp, regs->tp, regs->t0);
|
|
pr_cont(" t1 : " REG_FMT " t2 : " REG_FMT " s0 : " REG_FMT "\n",
|
|
regs->t1, regs->t2, regs->s0);
|
|
pr_cont(" s1 : " REG_FMT " a0 : " REG_FMT " a1 : " REG_FMT "\n",
|
|
regs->s1, regs->a0, regs->a1);
|
|
pr_cont(" a2 : " REG_FMT " a3 : " REG_FMT " a4 : " REG_FMT "\n",
|
|
regs->a2, regs->a3, regs->a4);
|
|
pr_cont(" a5 : " REG_FMT " a6 : " REG_FMT " a7 : " REG_FMT "\n",
|
|
regs->a5, regs->a6, regs->a7);
|
|
pr_cont(" s2 : " REG_FMT " s3 : " REG_FMT " s4 : " REG_FMT "\n",
|
|
regs->s2, regs->s3, regs->s4);
|
|
pr_cont(" s5 : " REG_FMT " s6 : " REG_FMT " s7 : " REG_FMT "\n",
|
|
regs->s5, regs->s6, regs->s7);
|
|
pr_cont(" s8 : " REG_FMT " s9 : " REG_FMT " s10: " REG_FMT "\n",
|
|
regs->s8, regs->s9, regs->s10);
|
|
pr_cont(" s11: " REG_FMT " t3 : " REG_FMT " t4 : " REG_FMT "\n",
|
|
regs->s11, regs->t3, regs->t4);
|
|
pr_cont(" t5 : " REG_FMT " t6 : " REG_FMT "\n",
|
|
regs->t5, regs->t6);
|
|
|
|
pr_cont("status: " REG_FMT " badaddr: " REG_FMT " cause: " REG_FMT "\n",
|
|
regs->status, regs->badaddr, regs->cause);
|
|
}
|
|
void show_regs(struct pt_regs *regs)
|
|
{
|
|
__show_regs(regs);
|
|
if (!user_mode(regs))
|
|
dump_backtrace(regs, NULL, KERN_DEFAULT);
|
|
}
|
|
|
|
unsigned long arch_align_stack(unsigned long sp)
|
|
{
|
|
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
|
|
sp -= get_random_u32_below(PAGE_SIZE);
|
|
return sp & ~0xf;
|
|
}
|
|
|
|
#ifdef CONFIG_COMPAT
|
|
static bool compat_mode_supported __read_mostly;
|
|
|
|
bool compat_elf_check_arch(Elf32_Ehdr *hdr)
|
|
{
|
|
return compat_mode_supported &&
|
|
hdr->e_machine == EM_RISCV &&
|
|
hdr->e_ident[EI_CLASS] == ELFCLASS32;
|
|
}
|
|
|
|
static int __init compat_mode_detect(void)
|
|
{
|
|
unsigned long tmp = csr_read(CSR_STATUS);
|
|
|
|
csr_write(CSR_STATUS, (tmp & ~SR_UXL) | SR_UXL_32);
|
|
compat_mode_supported =
|
|
(csr_read(CSR_STATUS) & SR_UXL) == SR_UXL_32;
|
|
|
|
csr_write(CSR_STATUS, tmp);
|
|
|
|
pr_info("riscv: ELF compat mode %s",
|
|
compat_mode_supported ? "supported" : "unsupported");
|
|
|
|
return 0;
|
|
}
|
|
early_initcall(compat_mode_detect);
|
|
#endif
|
|
|
|
void start_thread(struct pt_regs *regs, unsigned long pc,
|
|
unsigned long sp)
|
|
{
|
|
regs->status = SR_PIE;
|
|
if (has_fpu()) {
|
|
regs->status |= SR_FS_INITIAL;
|
|
/*
|
|
* Restore the initial value to the FP register
|
|
* before starting the user program.
|
|
*/
|
|
fstate_restore(current, regs);
|
|
}
|
|
regs->epc = pc;
|
|
regs->sp = sp;
|
|
|
|
#ifdef CONFIG_64BIT
|
|
regs->status &= ~SR_UXL;
|
|
|
|
if (is_compat_task())
|
|
regs->status |= SR_UXL_32;
|
|
else
|
|
regs->status |= SR_UXL_64;
|
|
#endif
|
|
}
|
|
|
|
void flush_thread(void)
|
|
{
|
|
#ifdef CONFIG_FPU
|
|
/*
|
|
* Reset FPU state and context
|
|
* frm: round to nearest, ties to even (IEEE default)
|
|
* fflags: accrued exceptions cleared
|
|
*/
|
|
fstate_off(current, task_pt_regs(current));
|
|
memset(¤t->thread.fstate, 0, sizeof(current->thread.fstate));
|
|
#endif
|
|
#ifdef CONFIG_RISCV_ISA_V
|
|
/* Reset vector state */
|
|
riscv_v_vstate_ctrl_init(current);
|
|
riscv_v_vstate_off(task_pt_regs(current));
|
|
kfree(current->thread.vstate.datap);
|
|
memset(¤t->thread.vstate, 0, sizeof(struct __riscv_v_ext_state));
|
|
clear_tsk_thread_flag(current, TIF_RISCV_V_DEFER_RESTORE);
|
|
#endif
|
|
}
|
|
|
|
void arch_release_task_struct(struct task_struct *tsk)
|
|
{
|
|
/* Free the vector context of datap. */
|
|
if (has_vector())
|
|
riscv_v_thread_free(tsk);
|
|
}
|
|
|
|
int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src)
|
|
{
|
|
fstate_save(src, task_pt_regs(src));
|
|
*dst = *src;
|
|
/* clear entire V context, including datap for a new task */
|
|
memset(&dst->thread.vstate, 0, sizeof(struct __riscv_v_ext_state));
|
|
memset(&dst->thread.kernel_vstate, 0, sizeof(struct __riscv_v_ext_state));
|
|
clear_tsk_thread_flag(dst, TIF_RISCV_V_DEFER_RESTORE);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int copy_thread(struct task_struct *p, const struct kernel_clone_args *args)
|
|
{
|
|
unsigned long clone_flags = args->flags;
|
|
unsigned long usp = args->stack;
|
|
unsigned long tls = args->tls;
|
|
struct pt_regs *childregs = task_pt_regs(p);
|
|
|
|
memset(&p->thread.s, 0, sizeof(p->thread.s));
|
|
|
|
/* p->thread holds context to be restored by __switch_to() */
|
|
if (unlikely(args->fn)) {
|
|
/* Kernel thread */
|
|
memset(childregs, 0, sizeof(struct pt_regs));
|
|
/* Supervisor/Machine, irqs on: */
|
|
childregs->status = SR_PP | SR_PIE;
|
|
|
|
p->thread.s[0] = (unsigned long)args->fn;
|
|
p->thread.s[1] = (unsigned long)args->fn_arg;
|
|
} else {
|
|
*childregs = *(current_pt_regs());
|
|
/* Turn off status.VS */
|
|
riscv_v_vstate_off(childregs);
|
|
if (usp) /* User fork */
|
|
childregs->sp = usp;
|
|
if (clone_flags & CLONE_SETTLS)
|
|
childregs->tp = tls;
|
|
childregs->a0 = 0; /* Return value of fork() */
|
|
p->thread.s[0] = 0;
|
|
}
|
|
p->thread.riscv_v_flags = 0;
|
|
if (has_vector())
|
|
riscv_v_thread_alloc(p);
|
|
p->thread.ra = (unsigned long)ret_from_fork;
|
|
p->thread.sp = (unsigned long)childregs; /* kernel sp */
|
|
return 0;
|
|
}
|
|
|
|
void __init arch_task_cache_init(void)
|
|
{
|
|
riscv_v_setup_ctx_cache();
|
|
}
|