mirror of
https://github.com/torvalds/linux.git
synced 2024-11-08 21:21:47 +00:00
fb481dd56a
Impact: performance optimization I did some rebenchmarking with modern compilers and dropping -funroll-loops makes the function consistently go faster by a few percent. So drop that flag. Thanks to Richard Guenther for a hint. Signed-off-by: Andi Kleen <ak@linux.intel.com> Signed-off-by: H. Peter Anvin <hpa@zytor.com>
25 lines
668 B
Makefile
25 lines
668 B
Makefile
#
|
|
# Makefile for x86 specific library files.
|
|
#
|
|
|
|
obj-$(CONFIG_SMP) := msr-on-cpu.o
|
|
|
|
lib-y := delay.o
|
|
lib-y += thunk_$(BITS).o
|
|
lib-y += usercopy_$(BITS).o getuser.o putuser.o
|
|
lib-y += memcpy_$(BITS).o
|
|
|
|
ifeq ($(CONFIG_X86_32),y)
|
|
lib-y += checksum_32.o
|
|
lib-y += strstr_32.o
|
|
lib-y += semaphore_32.o string_32.o
|
|
|
|
lib-$(CONFIG_X86_USE_3DNOW) += mmx_32.o
|
|
else
|
|
obj-y += io_64.o iomap_copy_64.o
|
|
lib-y += csum-partial_64.o csum-copy_64.o csum-wrappers_64.o
|
|
lib-y += thunk_64.o clear_page_64.o copy_page_64.o
|
|
lib-y += memmove_64.o memset_64.o
|
|
lib-y += copy_user_64.o rwlock_64.o copy_user_nocache_64.o
|
|
endif
|