forked from Minki/linux
fa48e6f780
The tlb maintainence functions: __cpu_flush_user_tlb_range and __cpu_flush_kern_tlb_range do not take into consideration the page granule when looping through the address range, and repeatedly flush tlb entries for the same page when operating with 64K pages. This patch re-works the logic s.t. we instead advance the loop by 1 << (PAGE_SHIFT - 12), so avoid repeating ourselves. Also the routines have been converted from assembler to static inline functions to aid with legibility and potential compiler optimisations. The isb() has been removed from flush_tlb_kernel_range(.) as it is only needed when changing the execute permission of a mapping. If one needs to set an area of the kernel as execute/non-execute an isb() must be inserted after the call to flush_tlb_kernel_range. Cc: Laura Abbott <lauraa@codeaurora.org> Signed-off-by: Steve Capper <steve.capper@linaro.org> Acked-by: Will Deacon <will.deacon@arm.com> Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
6 lines
195 B
Makefile
6 lines
195 B
Makefile
obj-y := dma-mapping.o extable.o fault.o init.o \
|
|
cache.o copypage.o flush.o \
|
|
ioremap.o mmap.o pgd.o mmu.o \
|
|
context.o proc.o
|
|
obj-$(CONFIG_HUGETLB_PAGE) += hugetlbpage.o
|