forked from Minki/linux
aed7eb8367
With ARM64_64K_PAGES and RANDOMIZE_TEXT_OFFSET enabled, we hit the following issue on the boot: kernel BUG at arch/arm64/mm/mmu.c:480! Internal error: Oops - BUG: 0 [#1] PREEMPT SMP Modules linked in: CPU: 0 PID: 0 Comm: swapper Not tainted 4.6.0 #310 Hardware name: ARM Juno development board (r2) (DT) task: ffff000008d58a80 ti: ffff000008d30000 task.ti: ffff000008d30000 PC is at map_kernel_segment+0x44/0xb0 LR is at paging_init+0x84/0x5b0 pc : [<ffff000008c450b4>] lr : [<ffff000008c451a4>] pstate: 600002c5 Call trace: [<ffff000008c450b4>] map_kernel_segment+0x44/0xb0 [<ffff000008c451a4>] paging_init+0x84/0x5b0 [<ffff000008c42728>] setup_arch+0x198/0x534 [<ffff000008c40848>] start_kernel+0x70/0x388 [<ffff000008c401bc>] __primary_switched+0x30/0x74 Commit7eb90f2ff7
("arm64: cover the .head.text section in the .text segment mapping") removed the alignment between the .head.text and .text sections, and used the _text rather than the _stext interval for mapping the .text segment. Prior to this commit _stext was always section aligned and didn't cause any issue even when RANDOMIZE_TEXT_OFFSET was enabled. Since that alignment has been removed and _text is used to map the .text segment, we need ensure _text is always page aligned when RANDOMIZE_TEXT_OFFSET is enabled. This patch adds logic to TEXT_OFFSET fuzzing to ensure that the offset is always aligned to the kernel page size. To ensure this, we rely on the PAGE_SHIFT being available via Kconfig. Signed-off-by: Mark Rutland <mark.rutland@arm.com> Reported-by: Sudeep Holla <sudeep.holla@arm.com> Cc: Ard Biesheuvel <ard.biesheuvel@linaro.org> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Will Deacon <will.deacon@arm.com> Fixes:7eb90f2ff7
("arm64: cover the .head.text section in the .text segment mapping") Signed-off-by: Will Deacon <will.deacon@arm.com>
135 lines
3.8 KiB
Makefile
135 lines
3.8 KiB
Makefile
#
|
|
# arch/arm64/Makefile
|
|
#
|
|
# This file is included by the global makefile so that you can add your own
|
|
# architecture-specific flags and dependencies.
|
|
#
|
|
# This file is subject to the terms and conditions of the GNU General Public
|
|
# License. See the file "COPYING" in the main directory of this archive
|
|
# for more details.
|
|
#
|
|
# Copyright (C) 1995-2001 by Russell King
|
|
|
|
LDFLAGS_vmlinux :=-p --no-undefined -X
|
|
CPPFLAGS_vmlinux.lds = -DTEXT_OFFSET=$(TEXT_OFFSET)
|
|
OBJCOPYFLAGS :=-O binary -R .note -R .note.gnu.build-id -R .comment -S
|
|
GZFLAGS :=-9
|
|
|
|
ifneq ($(CONFIG_RELOCATABLE),)
|
|
LDFLAGS_vmlinux += -pie
|
|
endif
|
|
|
|
KBUILD_DEFCONFIG := defconfig
|
|
|
|
# Check for binutils support for specific extensions
|
|
lseinstr := $(call as-instr,.arch_extension lse,-DCONFIG_AS_LSE=1)
|
|
|
|
ifeq ($(CONFIG_ARM64_LSE_ATOMICS), y)
|
|
ifeq ($(lseinstr),)
|
|
$(warning LSE atomics not supported by binutils)
|
|
endif
|
|
endif
|
|
|
|
KBUILD_CFLAGS += -mgeneral-regs-only $(lseinstr)
|
|
KBUILD_CFLAGS += -fno-asynchronous-unwind-tables
|
|
KBUILD_CFLAGS += $(call cc-option, -mpc-relative-literal-loads)
|
|
KBUILD_AFLAGS += $(lseinstr)
|
|
|
|
ifeq ($(CONFIG_CPU_BIG_ENDIAN), y)
|
|
KBUILD_CPPFLAGS += -mbig-endian
|
|
AS += -EB
|
|
LD += -EB
|
|
else
|
|
KBUILD_CPPFLAGS += -mlittle-endian
|
|
AS += -EL
|
|
LD += -EL
|
|
endif
|
|
|
|
CHECKFLAGS += -D__aarch64__
|
|
|
|
ifeq ($(CONFIG_ARM64_MODULE_CMODEL_LARGE), y)
|
|
KBUILD_CFLAGS_MODULE += -mcmodel=large
|
|
endif
|
|
|
|
ifeq ($(CONFIG_ARM64_MODULE_PLTS),y)
|
|
KBUILD_LDFLAGS_MODULE += -T $(srctree)/arch/arm64/kernel/module.lds
|
|
endif
|
|
|
|
# Default value
|
|
head-y := arch/arm64/kernel/head.o
|
|
|
|
# The byte offset of the kernel image in RAM from the start of RAM.
|
|
ifeq ($(CONFIG_ARM64_RANDOMIZE_TEXT_OFFSET), y)
|
|
TEXT_OFFSET := $(shell awk "BEGIN {srand(); printf \"0x%06x\n\", \
|
|
int(2 * 1024 * 1024 / (2 ^ $(CONFIG_ARM64_PAGE_SHIFT)) * \
|
|
rand()) * (2 ^ $(CONFIG_ARM64_PAGE_SHIFT))}")
|
|
else
|
|
TEXT_OFFSET := 0x00080000
|
|
endif
|
|
|
|
# KASAN_SHADOW_OFFSET = VA_START + (1 << (VA_BITS - 3)) - (1 << 61)
|
|
# in 32-bit arithmetic
|
|
KASAN_SHADOW_OFFSET := $(shell printf "0x%08x00000000\n" $$(( \
|
|
(0xffffffff & (-1 << ($(CONFIG_ARM64_VA_BITS) - 32))) \
|
|
+ (1 << ($(CONFIG_ARM64_VA_BITS) - 32 - 3)) \
|
|
- (1 << (64 - 32 - 3)) )) )
|
|
|
|
export TEXT_OFFSET GZFLAGS
|
|
|
|
core-y += arch/arm64/kernel/ arch/arm64/mm/
|
|
core-$(CONFIG_NET) += arch/arm64/net/
|
|
core-$(CONFIG_KVM) += arch/arm64/kvm/
|
|
core-$(CONFIG_XEN) += arch/arm64/xen/
|
|
core-$(CONFIG_CRYPTO) += arch/arm64/crypto/
|
|
libs-y := arch/arm64/lib/ $(libs-y)
|
|
core-$(CONFIG_EFI_STUB) += $(objtree)/drivers/firmware/efi/libstub/lib.a
|
|
|
|
# Default target when executing plain make
|
|
KBUILD_IMAGE := Image.gz
|
|
KBUILD_DTBS := dtbs
|
|
|
|
all: $(KBUILD_IMAGE) $(KBUILD_DTBS)
|
|
|
|
boot := arch/arm64/boot
|
|
|
|
Image: vmlinux
|
|
$(Q)$(MAKE) $(build)=$(boot) $(boot)/$@
|
|
|
|
Image.%: vmlinux
|
|
$(Q)$(MAKE) $(build)=$(boot) $(boot)/$@
|
|
|
|
zinstall install:
|
|
$(Q)$(MAKE) $(build)=$(boot) $@
|
|
|
|
%.dtb: scripts
|
|
$(Q)$(MAKE) $(build)=$(boot)/dts $(boot)/dts/$@
|
|
|
|
PHONY += dtbs dtbs_install
|
|
|
|
dtbs: prepare scripts
|
|
$(Q)$(MAKE) $(build)=$(boot)/dts
|
|
|
|
dtbs_install:
|
|
$(Q)$(MAKE) $(dtbinst)=$(boot)/dts
|
|
|
|
PHONY += vdso_install
|
|
vdso_install:
|
|
$(Q)$(MAKE) $(build)=arch/arm64/kernel/vdso $@
|
|
|
|
# We use MRPROPER_FILES and CLEAN_FILES now
|
|
archclean:
|
|
$(Q)$(MAKE) $(clean)=$(boot)
|
|
$(Q)$(MAKE) $(clean)=$(boot)/dts
|
|
|
|
define archhelp
|
|
echo '* Image.gz - Compressed kernel image (arch/$(ARCH)/boot/Image.gz)'
|
|
echo ' Image - Uncompressed kernel image (arch/$(ARCH)/boot/Image)'
|
|
echo '* dtbs - Build device tree blobs for enabled boards'
|
|
echo ' dtbs_install - Install dtbs to $(INSTALL_DTBS_PATH)'
|
|
echo ' install - Install uncompressed kernel'
|
|
echo ' zinstall - Install compressed kernel'
|
|
echo ' Install using (your) ~/bin/installkernel or'
|
|
echo ' (distribution) /sbin/installkernel or'
|
|
echo ' install to $$(INSTALL_PATH) and run lilo'
|
|
endef
|