forked from Minki/linux
f94909ceb1
Replace all ret/retq instructions with RET in preparation of making RET a macro. Since AS is case insensitive it's a big no-op without RET defined. find arch/x86/ -name \*.S | while read file do sed -i 's/\<ret[q]*\>/RET/' $file done Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Signed-off-by: Borislav Petkov <bp@suse.de> Link: https://lore.kernel.org/r/20211204134907.905503893@infradead.org
257 lines
7.0 KiB
ArmAsm
257 lines
7.0 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0 OR MIT */
|
|
/*
|
|
* Copyright (C) 2015-2019 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved.
|
|
* Copyright (C) 2017-2019 Samuel Neves <sneves@dei.uc.pt>. All Rights Reserved.
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
|
|
.section .rodata.cst32.BLAKE2S_IV, "aM", @progbits, 32
|
|
.align 32
|
|
IV: .octa 0xA54FF53A3C6EF372BB67AE856A09E667
|
|
.octa 0x5BE0CD191F83D9AB9B05688C510E527F
|
|
.section .rodata.cst16.ROT16, "aM", @progbits, 16
|
|
.align 16
|
|
ROT16: .octa 0x0D0C0F0E09080B0A0504070601000302
|
|
.section .rodata.cst16.ROR328, "aM", @progbits, 16
|
|
.align 16
|
|
ROR328: .octa 0x0C0F0E0D080B0A090407060500030201
|
|
.section .rodata.cst64.BLAKE2S_SIGMA, "aM", @progbits, 160
|
|
.align 64
|
|
SIGMA:
|
|
.byte 0, 2, 4, 6, 1, 3, 5, 7, 14, 8, 10, 12, 15, 9, 11, 13
|
|
.byte 14, 4, 9, 13, 10, 8, 15, 6, 5, 1, 0, 11, 3, 12, 2, 7
|
|
.byte 11, 12, 5, 15, 8, 0, 2, 13, 9, 10, 3, 7, 4, 14, 6, 1
|
|
.byte 7, 3, 13, 11, 9, 1, 12, 14, 15, 2, 5, 4, 8, 6, 10, 0
|
|
.byte 9, 5, 2, 10, 0, 7, 4, 15, 3, 14, 11, 6, 13, 1, 12, 8
|
|
.byte 2, 6, 0, 8, 12, 10, 11, 3, 1, 4, 7, 15, 9, 13, 5, 14
|
|
.byte 12, 1, 14, 4, 5, 15, 13, 10, 8, 0, 6, 9, 11, 7, 3, 2
|
|
.byte 13, 7, 12, 3, 11, 14, 1, 9, 2, 5, 15, 8, 10, 0, 4, 6
|
|
.byte 6, 14, 11, 0, 15, 9, 3, 8, 10, 12, 13, 1, 5, 2, 7, 4
|
|
.byte 10, 8, 7, 1, 2, 4, 6, 5, 13, 15, 9, 3, 0, 11, 14, 12
|
|
#ifdef CONFIG_AS_AVX512
|
|
.section .rodata.cst64.BLAKE2S_SIGMA2, "aM", @progbits, 640
|
|
.align 64
|
|
SIGMA2:
|
|
.long 0, 2, 4, 6, 1, 3, 5, 7, 14, 8, 10, 12, 15, 9, 11, 13
|
|
.long 8, 2, 13, 15, 10, 9, 12, 3, 6, 4, 0, 14, 5, 11, 1, 7
|
|
.long 11, 13, 8, 6, 5, 10, 14, 3, 2, 4, 12, 15, 1, 0, 7, 9
|
|
.long 11, 10, 7, 0, 8, 15, 1, 13, 3, 6, 2, 12, 4, 14, 9, 5
|
|
.long 4, 10, 9, 14, 15, 0, 11, 8, 1, 7, 3, 13, 2, 5, 6, 12
|
|
.long 2, 11, 4, 15, 14, 3, 10, 8, 13, 6, 5, 7, 0, 12, 1, 9
|
|
.long 4, 8, 15, 9, 14, 11, 13, 5, 3, 2, 1, 12, 6, 10, 7, 0
|
|
.long 6, 13, 0, 14, 12, 2, 1, 11, 15, 4, 5, 8, 7, 9, 3, 10
|
|
.long 15, 5, 4, 13, 10, 7, 3, 11, 12, 2, 0, 6, 9, 8, 1, 14
|
|
.long 8, 7, 14, 11, 13, 15, 0, 12, 10, 4, 5, 6, 3, 2, 1, 9
|
|
#endif /* CONFIG_AS_AVX512 */
|
|
|
|
.text
|
|
SYM_FUNC_START(blake2s_compress_ssse3)
|
|
testq %rdx,%rdx
|
|
je .Lendofloop
|
|
movdqu (%rdi),%xmm0
|
|
movdqu 0x10(%rdi),%xmm1
|
|
movdqa ROT16(%rip),%xmm12
|
|
movdqa ROR328(%rip),%xmm13
|
|
movdqu 0x20(%rdi),%xmm14
|
|
movq %rcx,%xmm15
|
|
leaq SIGMA+0xa0(%rip),%r8
|
|
jmp .Lbeginofloop
|
|
.align 32
|
|
.Lbeginofloop:
|
|
movdqa %xmm0,%xmm10
|
|
movdqa %xmm1,%xmm11
|
|
paddq %xmm15,%xmm14
|
|
movdqa IV(%rip),%xmm2
|
|
movdqa %xmm14,%xmm3
|
|
pxor IV+0x10(%rip),%xmm3
|
|
leaq SIGMA(%rip),%rcx
|
|
.Lroundloop:
|
|
movzbl (%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm4
|
|
movzbl 0x1(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm5
|
|
movzbl 0x2(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm6
|
|
movzbl 0x3(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm7
|
|
punpckldq %xmm5,%xmm4
|
|
punpckldq %xmm7,%xmm6
|
|
punpcklqdq %xmm6,%xmm4
|
|
paddd %xmm4,%xmm0
|
|
paddd %xmm1,%xmm0
|
|
pxor %xmm0,%xmm3
|
|
pshufb %xmm12,%xmm3
|
|
paddd %xmm3,%xmm2
|
|
pxor %xmm2,%xmm1
|
|
movdqa %xmm1,%xmm8
|
|
psrld $0xc,%xmm1
|
|
pslld $0x14,%xmm8
|
|
por %xmm8,%xmm1
|
|
movzbl 0x4(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm5
|
|
movzbl 0x5(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm6
|
|
movzbl 0x6(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm7
|
|
movzbl 0x7(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm4
|
|
punpckldq %xmm6,%xmm5
|
|
punpckldq %xmm4,%xmm7
|
|
punpcklqdq %xmm7,%xmm5
|
|
paddd %xmm5,%xmm0
|
|
paddd %xmm1,%xmm0
|
|
pxor %xmm0,%xmm3
|
|
pshufb %xmm13,%xmm3
|
|
paddd %xmm3,%xmm2
|
|
pxor %xmm2,%xmm1
|
|
movdqa %xmm1,%xmm8
|
|
psrld $0x7,%xmm1
|
|
pslld $0x19,%xmm8
|
|
por %xmm8,%xmm1
|
|
pshufd $0x93,%xmm0,%xmm0
|
|
pshufd $0x4e,%xmm3,%xmm3
|
|
pshufd $0x39,%xmm2,%xmm2
|
|
movzbl 0x8(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm6
|
|
movzbl 0x9(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm7
|
|
movzbl 0xa(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm4
|
|
movzbl 0xb(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm5
|
|
punpckldq %xmm7,%xmm6
|
|
punpckldq %xmm5,%xmm4
|
|
punpcklqdq %xmm4,%xmm6
|
|
paddd %xmm6,%xmm0
|
|
paddd %xmm1,%xmm0
|
|
pxor %xmm0,%xmm3
|
|
pshufb %xmm12,%xmm3
|
|
paddd %xmm3,%xmm2
|
|
pxor %xmm2,%xmm1
|
|
movdqa %xmm1,%xmm8
|
|
psrld $0xc,%xmm1
|
|
pslld $0x14,%xmm8
|
|
por %xmm8,%xmm1
|
|
movzbl 0xc(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm7
|
|
movzbl 0xd(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm4
|
|
movzbl 0xe(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm5
|
|
movzbl 0xf(%rcx),%eax
|
|
movd (%rsi,%rax,4),%xmm6
|
|
punpckldq %xmm4,%xmm7
|
|
punpckldq %xmm6,%xmm5
|
|
punpcklqdq %xmm5,%xmm7
|
|
paddd %xmm7,%xmm0
|
|
paddd %xmm1,%xmm0
|
|
pxor %xmm0,%xmm3
|
|
pshufb %xmm13,%xmm3
|
|
paddd %xmm3,%xmm2
|
|
pxor %xmm2,%xmm1
|
|
movdqa %xmm1,%xmm8
|
|
psrld $0x7,%xmm1
|
|
pslld $0x19,%xmm8
|
|
por %xmm8,%xmm1
|
|
pshufd $0x39,%xmm0,%xmm0
|
|
pshufd $0x4e,%xmm3,%xmm3
|
|
pshufd $0x93,%xmm2,%xmm2
|
|
addq $0x10,%rcx
|
|
cmpq %r8,%rcx
|
|
jnz .Lroundloop
|
|
pxor %xmm2,%xmm0
|
|
pxor %xmm3,%xmm1
|
|
pxor %xmm10,%xmm0
|
|
pxor %xmm11,%xmm1
|
|
addq $0x40,%rsi
|
|
decq %rdx
|
|
jnz .Lbeginofloop
|
|
movdqu %xmm0,(%rdi)
|
|
movdqu %xmm1,0x10(%rdi)
|
|
movdqu %xmm14,0x20(%rdi)
|
|
.Lendofloop:
|
|
RET
|
|
SYM_FUNC_END(blake2s_compress_ssse3)
|
|
|
|
#ifdef CONFIG_AS_AVX512
|
|
SYM_FUNC_START(blake2s_compress_avx512)
|
|
vmovdqu (%rdi),%xmm0
|
|
vmovdqu 0x10(%rdi),%xmm1
|
|
vmovdqu 0x20(%rdi),%xmm4
|
|
vmovq %rcx,%xmm5
|
|
vmovdqa IV(%rip),%xmm14
|
|
vmovdqa IV+16(%rip),%xmm15
|
|
jmp .Lblake2s_compress_avx512_mainloop
|
|
.align 32
|
|
.Lblake2s_compress_avx512_mainloop:
|
|
vmovdqa %xmm0,%xmm10
|
|
vmovdqa %xmm1,%xmm11
|
|
vpaddq %xmm5,%xmm4,%xmm4
|
|
vmovdqa %xmm14,%xmm2
|
|
vpxor %xmm15,%xmm4,%xmm3
|
|
vmovdqu (%rsi),%ymm6
|
|
vmovdqu 0x20(%rsi),%ymm7
|
|
addq $0x40,%rsi
|
|
leaq SIGMA2(%rip),%rax
|
|
movb $0xa,%cl
|
|
.Lblake2s_compress_avx512_roundloop:
|
|
addq $0x40,%rax
|
|
vmovdqa -0x40(%rax),%ymm8
|
|
vmovdqa -0x20(%rax),%ymm9
|
|
vpermi2d %ymm7,%ymm6,%ymm8
|
|
vpermi2d %ymm7,%ymm6,%ymm9
|
|
vmovdqa %ymm8,%ymm6
|
|
vmovdqa %ymm9,%ymm7
|
|
vpaddd %xmm8,%xmm0,%xmm0
|
|
vpaddd %xmm1,%xmm0,%xmm0
|
|
vpxor %xmm0,%xmm3,%xmm3
|
|
vprord $0x10,%xmm3,%xmm3
|
|
vpaddd %xmm3,%xmm2,%xmm2
|
|
vpxor %xmm2,%xmm1,%xmm1
|
|
vprord $0xc,%xmm1,%xmm1
|
|
vextracti128 $0x1,%ymm8,%xmm8
|
|
vpaddd %xmm8,%xmm0,%xmm0
|
|
vpaddd %xmm1,%xmm0,%xmm0
|
|
vpxor %xmm0,%xmm3,%xmm3
|
|
vprord $0x8,%xmm3,%xmm3
|
|
vpaddd %xmm3,%xmm2,%xmm2
|
|
vpxor %xmm2,%xmm1,%xmm1
|
|
vprord $0x7,%xmm1,%xmm1
|
|
vpshufd $0x93,%xmm0,%xmm0
|
|
vpshufd $0x4e,%xmm3,%xmm3
|
|
vpshufd $0x39,%xmm2,%xmm2
|
|
vpaddd %xmm9,%xmm0,%xmm0
|
|
vpaddd %xmm1,%xmm0,%xmm0
|
|
vpxor %xmm0,%xmm3,%xmm3
|
|
vprord $0x10,%xmm3,%xmm3
|
|
vpaddd %xmm3,%xmm2,%xmm2
|
|
vpxor %xmm2,%xmm1,%xmm1
|
|
vprord $0xc,%xmm1,%xmm1
|
|
vextracti128 $0x1,%ymm9,%xmm9
|
|
vpaddd %xmm9,%xmm0,%xmm0
|
|
vpaddd %xmm1,%xmm0,%xmm0
|
|
vpxor %xmm0,%xmm3,%xmm3
|
|
vprord $0x8,%xmm3,%xmm3
|
|
vpaddd %xmm3,%xmm2,%xmm2
|
|
vpxor %xmm2,%xmm1,%xmm1
|
|
vprord $0x7,%xmm1,%xmm1
|
|
vpshufd $0x39,%xmm0,%xmm0
|
|
vpshufd $0x4e,%xmm3,%xmm3
|
|
vpshufd $0x93,%xmm2,%xmm2
|
|
decb %cl
|
|
jne .Lblake2s_compress_avx512_roundloop
|
|
vpxor %xmm10,%xmm0,%xmm0
|
|
vpxor %xmm11,%xmm1,%xmm1
|
|
vpxor %xmm2,%xmm0,%xmm0
|
|
vpxor %xmm3,%xmm1,%xmm1
|
|
decq %rdx
|
|
jne .Lblake2s_compress_avx512_mainloop
|
|
vmovdqu %xmm0,(%rdi)
|
|
vmovdqu %xmm1,0x10(%rdi)
|
|
vmovdqu %xmm4,0x20(%rdi)
|
|
vzeroupper
|
|
RET
|
|
SYM_FUNC_END(blake2s_compress_avx512)
|
|
#endif /* CONFIG_AS_AVX512 */
|