A lot of asm-optimized routines in arch/x86/crypto/ keep its
constants in .data. This is wrong, they should be on .rodata.
Mnay of these constants are the same in different modules.
For example, 128-bit shuffle mask 0x000102030405060708090A0B0C0D0E0F
exists in at least half a dozen places.
There is a way to let linker merge them and use just one copy.
The rules are as follows: mergeable objects of different sizes
should not share sections. You can't put them all in one .rodata
section, they will lose "mergeability".
GCC puts its mergeable constants in ".rodata.cstSIZE" sections,
or ".rodata.cstSIZE.<object_name>" if -fdata-sections is used.
This patch does the same:
	.section .rodata.cst16.SHUF_MASK, "aM", @progbits, 16
It is important that all data in such section consists of
16-byte elements, not larger ones, and there are no implicit
use of one element from another.
When this is not the case, use non-mergeable section:
	.section .rodata[.VAR_NAME], "a", @progbits
This reduces .data by ~15 kbytes:
    text    data     bss     dec      hex filename
11097415 2705840 2630712 16433967  fac32f vmlinux-prev.o
11112095 2690672 2630712 16433479  fac147 vmlinux.o
Merged objects are visible in System.map:
ffffffff81a28810 r POLY
ffffffff81a28810 r POLY
ffffffff81a28820 r TWOONE
ffffffff81a28820 r TWOONE
ffffffff81a28830 r PSHUFFLE_BYTE_FLIP_MASK <- merged regardless of
ffffffff81a28830 r SHUF_MASK   <------------- the name difference
ffffffff81a28830 r SHUF_MASK
ffffffff81a28830 r SHUF_MASK
..
ffffffff81a28d00 r K512 <- merged three identical 640-byte tables
ffffffff81a28d00 r K512
ffffffff81a28d00 r K512
Use of object names in section name suffixes is not strictly necessary,
but might help if someday link stage will use garbage collection
to eliminate unused sections (ld --gc-sections).
Signed-off-by: Denys Vlasenko <dvlasenk@redhat.com>
CC: Herbert Xu <herbert@gondor.apana.org.au>
CC: Josh Poimboeuf <jpoimboe@redhat.com>
CC: Xiaodong Liu <xiaodong.liu@intel.com>
CC: Megha Dey <megha.dey@intel.com>
CC: linux-crypto@vger.kernel.org
CC: x86@kernel.org
CC: linux-kernel@vger.kernel.org
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
		
	
			
		
			
				
	
	
		
			215 lines
		
	
	
		
			6.1 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
			
		
		
	
	
			215 lines
		
	
	
		
			6.1 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
| /*
 | |
|  * Buffer submit code for multi buffer SHA256 algorithm
 | |
|  *
 | |
|  * This file is provided under a dual BSD/GPLv2 license.  When using or
 | |
|  * redistributing this file, you may do so under either license.
 | |
|  *
 | |
|  * GPL LICENSE SUMMARY
 | |
|  *
 | |
|  *  Copyright(c) 2016 Intel Corporation.
 | |
|  *
 | |
|  *  This program is free software; you can redistribute it and/or modify
 | |
|  *  it under the terms of version 2 of the GNU General Public License as
 | |
|  *  published by the Free Software Foundation.
 | |
|  *
 | |
|  *  This program is distributed in the hope that it will be useful, but
 | |
|  *  WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  *  General Public License for more details.
 | |
|  *
 | |
|  *  Contact Information:
 | |
|  *      Megha Dey <megha.dey@linux.intel.com>
 | |
|  *
 | |
|  *  BSD LICENSE
 | |
|  *
 | |
|  *  Copyright(c) 2016 Intel Corporation.
 | |
|  *
 | |
|  *  Redistribution and use in source and binary forms, with or without
 | |
|  *  modification, are permitted provided that the following conditions
 | |
|  *  are met:
 | |
|  *
 | |
|  *    * Redistributions of source code must retain the above copyright
 | |
|  *      notice, this list of conditions and the following disclaimer.
 | |
|  *    * Redistributions in binary form must reproduce the above copyright
 | |
|  *      notice, this list of conditions and the following disclaimer in
 | |
|  *      the documentation and/or other materials provided with the
 | |
|  *      distribution.
 | |
|  *    * Neither the name of Intel Corporation nor the names of its
 | |
|  *      contributors may be used to endorse or promote products derived
 | |
|  *      from this software without specific prior written permission.
 | |
|  *
 | |
|  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
 | |
|  *  "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
 | |
|  *  LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
 | |
|  *  A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
 | |
|  *  OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 | |
|  *  SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
 | |
|  *  LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 | |
|  *  DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 | |
|  *  THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 | |
|  *  (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
 | |
|  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 | |
|  */
 | |
| 
 | |
| #include <linux/linkage.h>
 | |
| #include <asm/frame.h>
 | |
| #include "sha256_mb_mgr_datastruct.S"
 | |
| 
 | |
| .extern sha256_x8_avx2
 | |
| 
 | |
| # LINUX register definitions
 | |
| arg1		= %rdi
 | |
| arg2		= %rsi
 | |
| size_offset	= %rcx
 | |
| tmp2		= %rcx
 | |
| extra_blocks	= %rdx
 | |
| 
 | |
| # Common definitions
 | |
| #define state	arg1
 | |
| #define job	%rsi
 | |
| #define len2	arg2
 | |
| #define p2	arg2
 | |
| 
 | |
| # idx must be a register not clobberred by sha1_x8_avx2
 | |
| idx		= %r8
 | |
| DWORD_idx	= %r8d
 | |
| last_len	= %r8
 | |
| 
 | |
| p		= %r11
 | |
| start_offset	= %r11
 | |
| 
 | |
| unused_lanes	= %rbx
 | |
| BYTE_unused_lanes = %bl
 | |
| 
 | |
| job_rax		= %rax
 | |
| len		= %rax
 | |
| DWORD_len	= %eax
 | |
| 
 | |
| lane		= %r12
 | |
| tmp3		= %r12
 | |
| 
 | |
| tmp		= %r9
 | |
| DWORD_tmp	= %r9d
 | |
| 
 | |
| lane_data	= %r10
 | |
| 
 | |
| # JOB* sha256_mb_mgr_submit_avx2(MB_MGR *state, JOB_SHA256 *job)
 | |
| # arg 1 : rcx : state
 | |
| # arg 2 : rdx : job
 | |
| ENTRY(sha256_mb_mgr_submit_avx2)
 | |
| 	FRAME_BEGIN
 | |
| 	push	%rbx
 | |
| 	push	%r12
 | |
| 
 | |
| 	mov	_unused_lanes(state), unused_lanes
 | |
| 	mov	unused_lanes, lane
 | |
| 	and	$0xF, lane
 | |
| 	shr	$4, unused_lanes
 | |
| 	imul	$_LANE_DATA_size, lane, lane_data
 | |
| 	movl	$STS_BEING_PROCESSED, _status(job)
 | |
| 	lea	_ldata(state, lane_data), lane_data
 | |
| 	mov	unused_lanes, _unused_lanes(state)
 | |
| 	movl	_len(job),  DWORD_len
 | |
| 
 | |
| 	mov	job, _job_in_lane(lane_data)
 | |
| 	shl	$4, len
 | |
| 	or	lane, len
 | |
| 
 | |
| 	movl	DWORD_len,  _lens(state , lane, 4)
 | |
| 
 | |
| 	# Load digest words from result_digest
 | |
| 	vmovdqu	_result_digest(job), %xmm0
 | |
| 	vmovdqu	_result_digest+1*16(job), %xmm1
 | |
| 	vmovd	%xmm0, _args_digest(state, lane, 4)
 | |
| 	vpextrd	$1, %xmm0, _args_digest+1*32(state , lane, 4)
 | |
| 	vpextrd	$2, %xmm0, _args_digest+2*32(state , lane, 4)
 | |
| 	vpextrd	$3, %xmm0, _args_digest+3*32(state , lane, 4)
 | |
| 	vmovd	%xmm1, _args_digest+4*32(state , lane, 4)
 | |
| 
 | |
| 	vpextrd	$1, %xmm1, _args_digest+5*32(state , lane, 4)
 | |
| 	vpextrd	$2, %xmm1, _args_digest+6*32(state , lane, 4)
 | |
| 	vpextrd	$3, %xmm1, _args_digest+7*32(state , lane, 4)
 | |
| 
 | |
| 	mov	_buffer(job), p
 | |
| 	mov	p, _args_data_ptr(state, lane, 8)
 | |
| 
 | |
| 	cmp	$0xF, unused_lanes
 | |
| 	jne	return_null
 | |
| 
 | |
| start_loop:
 | |
| 	# Find min length
 | |
| 	vmovdqa	_lens(state), %xmm0
 | |
| 	vmovdqa	_lens+1*16(state), %xmm1
 | |
| 
 | |
| 	vpminud	%xmm1, %xmm0, %xmm2		# xmm2 has {D,C,B,A}
 | |
| 	vpalignr $8, %xmm2, %xmm3, %xmm3	# xmm3 has {x,x,D,C}
 | |
| 	vpminud	%xmm3, %xmm2, %xmm2		# xmm2 has {x,x,E,F}
 | |
| 	vpalignr $4, %xmm2, %xmm3, %xmm3	# xmm3 has {x,x,x,E}
 | |
| 	vpminud	%xmm3, %xmm2, %xmm2		# xmm2 has min val in low dword
 | |
| 
 | |
| 	vmovd	%xmm2, DWORD_idx
 | |
| 	mov	idx, len2
 | |
| 	and	$0xF, idx
 | |
| 	shr	$4, len2
 | |
| 	jz	len_is_0
 | |
| 
 | |
| 	vpand	clear_low_nibble(%rip), %xmm2, %xmm2
 | |
| 	vpshufd	$0, %xmm2, %xmm2
 | |
| 
 | |
| 	vpsubd	%xmm2, %xmm0, %xmm0
 | |
| 	vpsubd	%xmm2, %xmm1, %xmm1
 | |
| 
 | |
| 	vmovdqa	%xmm0, _lens + 0*16(state)
 | |
| 	vmovdqa	%xmm1, _lens + 1*16(state)
 | |
| 
 | |
| 	# "state" and "args" are the same address, arg1
 | |
| 	# len is arg2
 | |
| 	call	sha256_x8_avx2
 | |
| 
 | |
| 	# state and idx are intact
 | |
| 
 | |
| len_is_0:
 | |
| 	# process completed job "idx"
 | |
| 	imul	$_LANE_DATA_size, idx, lane_data
 | |
| 	lea	_ldata(state, lane_data), lane_data
 | |
| 
 | |
| 	mov	_job_in_lane(lane_data), job_rax
 | |
| 	mov	_unused_lanes(state), unused_lanes
 | |
| 	movq	$0, _job_in_lane(lane_data)
 | |
| 	movl	$STS_COMPLETED, _status(job_rax)
 | |
| 	shl	$4, unused_lanes
 | |
| 	or	idx, unused_lanes
 | |
| 	mov	unused_lanes, _unused_lanes(state)
 | |
| 
 | |
| 	movl	$0xFFFFFFFF, _lens(state,idx,4)
 | |
| 
 | |
| 	vmovd	_args_digest(state, idx, 4), %xmm0
 | |
| 	vpinsrd	$1, _args_digest+1*32(state , idx, 4), %xmm0, %xmm0
 | |
| 	vpinsrd	$2, _args_digest+2*32(state , idx, 4), %xmm0, %xmm0
 | |
| 	vpinsrd	$3, _args_digest+3*32(state , idx, 4), %xmm0, %xmm0
 | |
| 	vmovd	_args_digest+4*32(state, idx, 4), %xmm1
 | |
| 
 | |
| 	vpinsrd	$1, _args_digest+5*32(state , idx, 4), %xmm1, %xmm1
 | |
| 	vpinsrd	$2, _args_digest+6*32(state , idx, 4), %xmm1, %xmm1
 | |
| 	vpinsrd	$3, _args_digest+7*32(state , idx, 4), %xmm1, %xmm1
 | |
| 
 | |
| 	vmovdqu	%xmm0, _result_digest(job_rax)
 | |
| 	vmovdqu	%xmm1, _result_digest+1*16(job_rax)
 | |
| 
 | |
| return:
 | |
| 	pop     %r12
 | |
|         pop     %rbx
 | |
|         FRAME_END
 | |
| 	ret
 | |
| 
 | |
| return_null:
 | |
| 	xor	job_rax, job_rax
 | |
| 	jmp	return
 | |
| 
 | |
| ENDPROC(sha256_mb_mgr_submit_avx2)
 | |
| 
 | |
| .section	.rodata.cst16.clear_low_nibble, "aM", @progbits, 16
 | |
| .align 16
 | |
| clear_low_nibble:
 | |
| 	.octa	0x000000000000000000000000FFFFFFF0
 |