mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 13:41:51 +00:00
47446d7cd4
aesbs_ecb_encrypt(), aesbs_ecb_decrypt(), aesbs_xts_encrypt(), and
aesbs_xts_decrypt() are called via indirect function calls. Therefore
they need to use SYM_TYPED_FUNC_START instead of SYM_FUNC_START to cause
their type hashes to be emitted when the kernel is built with
CONFIG_CFI_CLANG=y. Otherwise, the code crashes with a CFI failure if
the compiler doesn't happen to optimize out the indirect calls.
Fixes: c50d32859e
("arm64: Add types to indirect called assembly functions")
Cc: stable@vger.kernel.org
Signed-off-by: Eric Biggers <ebiggers@google.com>
Reviewed-by: Ard Biesheuvel <ardb@kernel.org>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
867 lines
22 KiB
ArmAsm
867 lines
22 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Bit sliced AES using NEON instructions
|
|
*
|
|
* Copyright (C) 2016 Linaro Ltd <ard.biesheuvel@linaro.org>
|
|
*/
|
|
|
|
/*
|
|
* The algorithm implemented here is described in detail by the paper
|
|
* 'Faster and Timing-Attack Resistant AES-GCM' by Emilia Kaesper and
|
|
* Peter Schwabe (https://eprint.iacr.org/2009/129.pdf)
|
|
*
|
|
* This implementation is based primarily on the OpenSSL implementation
|
|
* for 32-bit ARM written by Andy Polyakov <appro@openssl.org>
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
#include <linux/cfi_types.h>
|
|
#include <asm/assembler.h>
|
|
|
|
.text
|
|
|
|
rounds .req x11
|
|
bskey .req x12
|
|
|
|
.macro in_bs_ch, b0, b1, b2, b3, b4, b5, b6, b7
|
|
eor \b2, \b2, \b1
|
|
eor \b5, \b5, \b6
|
|
eor \b3, \b3, \b0
|
|
eor \b6, \b6, \b2
|
|
eor \b5, \b5, \b0
|
|
eor \b6, \b6, \b3
|
|
eor \b3, \b3, \b7
|
|
eor \b7, \b7, \b5
|
|
eor \b3, \b3, \b4
|
|
eor \b4, \b4, \b5
|
|
eor \b2, \b2, \b7
|
|
eor \b3, \b3, \b1
|
|
eor \b1, \b1, \b5
|
|
.endm
|
|
|
|
.macro out_bs_ch, b0, b1, b2, b3, b4, b5, b6, b7
|
|
eor \b0, \b0, \b6
|
|
eor \b1, \b1, \b4
|
|
eor \b4, \b4, \b6
|
|
eor \b2, \b2, \b0
|
|
eor \b6, \b6, \b1
|
|
eor \b1, \b1, \b5
|
|
eor \b5, \b5, \b3
|
|
eor \b3, \b3, \b7
|
|
eor \b7, \b7, \b5
|
|
eor \b2, \b2, \b5
|
|
eor \b4, \b4, \b7
|
|
.endm
|
|
|
|
.macro inv_in_bs_ch, b6, b1, b2, b4, b7, b0, b3, b5
|
|
eor \b1, \b1, \b7
|
|
eor \b4, \b4, \b7
|
|
eor \b7, \b7, \b5
|
|
eor \b1, \b1, \b3
|
|
eor \b2, \b2, \b5
|
|
eor \b3, \b3, \b7
|
|
eor \b6, \b6, \b1
|
|
eor \b2, \b2, \b0
|
|
eor \b5, \b5, \b3
|
|
eor \b4, \b4, \b6
|
|
eor \b0, \b0, \b6
|
|
eor \b1, \b1, \b4
|
|
.endm
|
|
|
|
.macro inv_out_bs_ch, b6, b5, b0, b3, b7, b1, b4, b2
|
|
eor \b1, \b1, \b5
|
|
eor \b2, \b2, \b7
|
|
eor \b3, \b3, \b1
|
|
eor \b4, \b4, \b5
|
|
eor \b7, \b7, \b5
|
|
eor \b3, \b3, \b4
|
|
eor \b5, \b5, \b0
|
|
eor \b3, \b3, \b7
|
|
eor \b6, \b6, \b2
|
|
eor \b2, \b2, \b1
|
|
eor \b6, \b6, \b3
|
|
eor \b3, \b3, \b0
|
|
eor \b5, \b5, \b6
|
|
.endm
|
|
|
|
.macro mul_gf4, x0, x1, y0, y1, t0, t1
|
|
eor \t0, \y0, \y1
|
|
and \t0, \t0, \x0
|
|
eor \x0, \x0, \x1
|
|
and \t1, \x1, \y0
|
|
and \x0, \x0, \y1
|
|
eor \x1, \t1, \t0
|
|
eor \x0, \x0, \t1
|
|
.endm
|
|
|
|
.macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1
|
|
eor \t0, \y0, \y1
|
|
eor \t1, \y2, \y3
|
|
and \t0, \t0, \x0
|
|
and \t1, \t1, \x2
|
|
eor \x0, \x0, \x1
|
|
eor \x2, \x2, \x3
|
|
and \x1, \x1, \y0
|
|
and \x3, \x3, \y2
|
|
and \x0, \x0, \y1
|
|
and \x2, \x2, \y3
|
|
eor \x1, \x1, \x0
|
|
eor \x2, \x2, \x3
|
|
eor \x0, \x0, \t0
|
|
eor \x3, \x3, \t1
|
|
.endm
|
|
|
|
.macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
y0, y1, y2, y3, t0, t1, t2, t3
|
|
eor \t0, \x0, \x2
|
|
eor \t1, \x1, \x3
|
|
mul_gf4 \x0, \x1, \y0, \y1, \t2, \t3
|
|
eor \y0, \y0, \y2
|
|
eor \y1, \y1, \y3
|
|
mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2
|
|
eor \x0, \x0, \t0
|
|
eor \x2, \x2, \t0
|
|
eor \x1, \x1, \t1
|
|
eor \x3, \x3, \t1
|
|
eor \t0, \x4, \x6
|
|
eor \t1, \x5, \x7
|
|
mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2
|
|
eor \y0, \y0, \y2
|
|
eor \y1, \y1, \y3
|
|
mul_gf4 \x4, \x5, \y0, \y1, \t2, \t3
|
|
eor \x4, \x4, \t0
|
|
eor \x6, \x6, \t0
|
|
eor \x5, \x5, \t1
|
|
eor \x7, \x7, \t1
|
|
.endm
|
|
|
|
.macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
t0, t1, t2, t3, s0, s1, s2, s3
|
|
eor \t3, \x4, \x6
|
|
eor \t0, \x5, \x7
|
|
eor \t1, \x1, \x3
|
|
eor \s1, \x7, \x6
|
|
eor \s0, \x0, \x2
|
|
eor \s3, \t3, \t0
|
|
orr \t2, \t0, \t1
|
|
and \s2, \t3, \s0
|
|
orr \t3, \t3, \s0
|
|
eor \s0, \s0, \t1
|
|
and \t0, \t0, \t1
|
|
eor \t1, \x3, \x2
|
|
and \s3, \s3, \s0
|
|
and \s1, \s1, \t1
|
|
eor \t1, \x4, \x5
|
|
eor \s0, \x1, \x0
|
|
eor \t3, \t3, \s1
|
|
eor \t2, \t2, \s1
|
|
and \s1, \t1, \s0
|
|
orr \t1, \t1, \s0
|
|
eor \t3, \t3, \s3
|
|
eor \t0, \t0, \s1
|
|
eor \t2, \t2, \s2
|
|
eor \t1, \t1, \s3
|
|
eor \t0, \t0, \s2
|
|
and \s0, \x7, \x3
|
|
eor \t1, \t1, \s2
|
|
and \s1, \x6, \x2
|
|
and \s2, \x5, \x1
|
|
orr \s3, \x4, \x0
|
|
eor \t3, \t3, \s0
|
|
eor \t1, \t1, \s2
|
|
eor \s0, \t0, \s3
|
|
eor \t2, \t2, \s1
|
|
and \s2, \t3, \t1
|
|
eor \s1, \t2, \s2
|
|
eor \s3, \s0, \s2
|
|
bsl \s1, \t1, \s0
|
|
not \t0, \s0
|
|
bsl \s0, \s1, \s3
|
|
bsl \t0, \s1, \s3
|
|
bsl \s3, \t3, \t2
|
|
eor \t3, \t3, \t2
|
|
and \s2, \s0, \s3
|
|
eor \t1, \t1, \t0
|
|
eor \s2, \s2, \t3
|
|
mul_gf16_2 \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \
|
|
\s3, \s2, \s1, \t1, \s0, \t0, \t2, \t3
|
|
.endm
|
|
|
|
.macro sbox, b0, b1, b2, b3, b4, b5, b6, b7, \
|
|
t0, t1, t2, t3, s0, s1, s2, s3
|
|
in_bs_ch \b0\().16b, \b1\().16b, \b2\().16b, \b3\().16b, \
|
|
\b4\().16b, \b5\().16b, \b6\().16b, \b7\().16b
|
|
inv_gf256 \b6\().16b, \b5\().16b, \b0\().16b, \b3\().16b, \
|
|
\b7\().16b, \b1\().16b, \b4\().16b, \b2\().16b, \
|
|
\t0\().16b, \t1\().16b, \t2\().16b, \t3\().16b, \
|
|
\s0\().16b, \s1\().16b, \s2\().16b, \s3\().16b
|
|
out_bs_ch \b7\().16b, \b1\().16b, \b4\().16b, \b2\().16b, \
|
|
\b6\().16b, \b5\().16b, \b0\().16b, \b3\().16b
|
|
.endm
|
|
|
|
.macro inv_sbox, b0, b1, b2, b3, b4, b5, b6, b7, \
|
|
t0, t1, t2, t3, s0, s1, s2, s3
|
|
inv_in_bs_ch \b0\().16b, \b1\().16b, \b2\().16b, \b3\().16b, \
|
|
\b4\().16b, \b5\().16b, \b6\().16b, \b7\().16b
|
|
inv_gf256 \b5\().16b, \b1\().16b, \b2\().16b, \b6\().16b, \
|
|
\b3\().16b, \b7\().16b, \b0\().16b, \b4\().16b, \
|
|
\t0\().16b, \t1\().16b, \t2\().16b, \t3\().16b, \
|
|
\s0\().16b, \s1\().16b, \s2\().16b, \s3\().16b
|
|
inv_out_bs_ch \b3\().16b, \b7\().16b, \b0\().16b, \b4\().16b, \
|
|
\b5\().16b, \b1\().16b, \b2\().16b, \b6\().16b
|
|
.endm
|
|
|
|
.macro enc_next_rk
|
|
ldp q16, q17, [bskey], #128
|
|
ldp q18, q19, [bskey, #-96]
|
|
ldp q20, q21, [bskey, #-64]
|
|
ldp q22, q23, [bskey, #-32]
|
|
.endm
|
|
|
|
.macro dec_next_rk
|
|
ldp q16, q17, [bskey, #-128]!
|
|
ldp q18, q19, [bskey, #32]
|
|
ldp q20, q21, [bskey, #64]
|
|
ldp q22, q23, [bskey, #96]
|
|
.endm
|
|
|
|
.macro add_round_key, x0, x1, x2, x3, x4, x5, x6, x7
|
|
eor \x0\().16b, \x0\().16b, v16.16b
|
|
eor \x1\().16b, \x1\().16b, v17.16b
|
|
eor \x2\().16b, \x2\().16b, v18.16b
|
|
eor \x3\().16b, \x3\().16b, v19.16b
|
|
eor \x4\().16b, \x4\().16b, v20.16b
|
|
eor \x5\().16b, \x5\().16b, v21.16b
|
|
eor \x6\().16b, \x6\().16b, v22.16b
|
|
eor \x7\().16b, \x7\().16b, v23.16b
|
|
.endm
|
|
|
|
.macro shift_rows, x0, x1, x2, x3, x4, x5, x6, x7, mask
|
|
tbl \x0\().16b, {\x0\().16b}, \mask\().16b
|
|
tbl \x1\().16b, {\x1\().16b}, \mask\().16b
|
|
tbl \x2\().16b, {\x2\().16b}, \mask\().16b
|
|
tbl \x3\().16b, {\x3\().16b}, \mask\().16b
|
|
tbl \x4\().16b, {\x4\().16b}, \mask\().16b
|
|
tbl \x5\().16b, {\x5\().16b}, \mask\().16b
|
|
tbl \x6\().16b, {\x6\().16b}, \mask\().16b
|
|
tbl \x7\().16b, {\x7\().16b}, \mask\().16b
|
|
.endm
|
|
|
|
.macro mix_cols, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
t0, t1, t2, t3, t4, t5, t6, t7, inv
|
|
ext \t0\().16b, \x0\().16b, \x0\().16b, #12
|
|
ext \t1\().16b, \x1\().16b, \x1\().16b, #12
|
|
eor \x0\().16b, \x0\().16b, \t0\().16b
|
|
ext \t2\().16b, \x2\().16b, \x2\().16b, #12
|
|
eor \x1\().16b, \x1\().16b, \t1\().16b
|
|
ext \t3\().16b, \x3\().16b, \x3\().16b, #12
|
|
eor \x2\().16b, \x2\().16b, \t2\().16b
|
|
ext \t4\().16b, \x4\().16b, \x4\().16b, #12
|
|
eor \x3\().16b, \x3\().16b, \t3\().16b
|
|
ext \t5\().16b, \x5\().16b, \x5\().16b, #12
|
|
eor \x4\().16b, \x4\().16b, \t4\().16b
|
|
ext \t6\().16b, \x6\().16b, \x6\().16b, #12
|
|
eor \x5\().16b, \x5\().16b, \t5\().16b
|
|
ext \t7\().16b, \x7\().16b, \x7\().16b, #12
|
|
eor \x6\().16b, \x6\().16b, \t6\().16b
|
|
eor \t1\().16b, \t1\().16b, \x0\().16b
|
|
eor \x7\().16b, \x7\().16b, \t7\().16b
|
|
ext \x0\().16b, \x0\().16b, \x0\().16b, #8
|
|
eor \t2\().16b, \t2\().16b, \x1\().16b
|
|
eor \t0\().16b, \t0\().16b, \x7\().16b
|
|
eor \t1\().16b, \t1\().16b, \x7\().16b
|
|
ext \x1\().16b, \x1\().16b, \x1\().16b, #8
|
|
eor \t5\().16b, \t5\().16b, \x4\().16b
|
|
eor \x0\().16b, \x0\().16b, \t0\().16b
|
|
eor \t6\().16b, \t6\().16b, \x5\().16b
|
|
eor \x1\().16b, \x1\().16b, \t1\().16b
|
|
ext \t0\().16b, \x4\().16b, \x4\().16b, #8
|
|
eor \t4\().16b, \t4\().16b, \x3\().16b
|
|
ext \t1\().16b, \x5\().16b, \x5\().16b, #8
|
|
eor \t7\().16b, \t7\().16b, \x6\().16b
|
|
ext \x4\().16b, \x3\().16b, \x3\().16b, #8
|
|
eor \t3\().16b, \t3\().16b, \x2\().16b
|
|
ext \x5\().16b, \x7\().16b, \x7\().16b, #8
|
|
eor \t4\().16b, \t4\().16b, \x7\().16b
|
|
ext \x3\().16b, \x6\().16b, \x6\().16b, #8
|
|
eor \t3\().16b, \t3\().16b, \x7\().16b
|
|
ext \x6\().16b, \x2\().16b, \x2\().16b, #8
|
|
eor \x7\().16b, \t1\().16b, \t5\().16b
|
|
.ifb \inv
|
|
eor \x2\().16b, \t0\().16b, \t4\().16b
|
|
eor \x4\().16b, \x4\().16b, \t3\().16b
|
|
eor \x5\().16b, \x5\().16b, \t7\().16b
|
|
eor \x3\().16b, \x3\().16b, \t6\().16b
|
|
eor \x6\().16b, \x6\().16b, \t2\().16b
|
|
.else
|
|
eor \t3\().16b, \t3\().16b, \x4\().16b
|
|
eor \x5\().16b, \x5\().16b, \t7\().16b
|
|
eor \x2\().16b, \x3\().16b, \t6\().16b
|
|
eor \x3\().16b, \t0\().16b, \t4\().16b
|
|
eor \x4\().16b, \x6\().16b, \t2\().16b
|
|
mov \x6\().16b, \t3\().16b
|
|
.endif
|
|
.endm
|
|
|
|
.macro inv_mix_cols, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
t0, t1, t2, t3, t4, t5, t6, t7
|
|
ext \t0\().16b, \x0\().16b, \x0\().16b, #8
|
|
ext \t6\().16b, \x6\().16b, \x6\().16b, #8
|
|
ext \t7\().16b, \x7\().16b, \x7\().16b, #8
|
|
eor \t0\().16b, \t0\().16b, \x0\().16b
|
|
ext \t1\().16b, \x1\().16b, \x1\().16b, #8
|
|
eor \t6\().16b, \t6\().16b, \x6\().16b
|
|
ext \t2\().16b, \x2\().16b, \x2\().16b, #8
|
|
eor \t7\().16b, \t7\().16b, \x7\().16b
|
|
ext \t3\().16b, \x3\().16b, \x3\().16b, #8
|
|
eor \t1\().16b, \t1\().16b, \x1\().16b
|
|
ext \t4\().16b, \x4\().16b, \x4\().16b, #8
|
|
eor \t2\().16b, \t2\().16b, \x2\().16b
|
|
ext \t5\().16b, \x5\().16b, \x5\().16b, #8
|
|
eor \t3\().16b, \t3\().16b, \x3\().16b
|
|
eor \t4\().16b, \t4\().16b, \x4\().16b
|
|
eor \t5\().16b, \t5\().16b, \x5\().16b
|
|
eor \x0\().16b, \x0\().16b, \t6\().16b
|
|
eor \x1\().16b, \x1\().16b, \t6\().16b
|
|
eor \x2\().16b, \x2\().16b, \t0\().16b
|
|
eor \x4\().16b, \x4\().16b, \t2\().16b
|
|
eor \x3\().16b, \x3\().16b, \t1\().16b
|
|
eor \x1\().16b, \x1\().16b, \t7\().16b
|
|
eor \x2\().16b, \x2\().16b, \t7\().16b
|
|
eor \x4\().16b, \x4\().16b, \t6\().16b
|
|
eor \x5\().16b, \x5\().16b, \t3\().16b
|
|
eor \x3\().16b, \x3\().16b, \t6\().16b
|
|
eor \x6\().16b, \x6\().16b, \t4\().16b
|
|
eor \x4\().16b, \x4\().16b, \t7\().16b
|
|
eor \x5\().16b, \x5\().16b, \t7\().16b
|
|
eor \x7\().16b, \x7\().16b, \t5\().16b
|
|
mix_cols \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \
|
|
\t0, \t1, \t2, \t3, \t4, \t5, \t6, \t7, 1
|
|
.endm
|
|
|
|
.macro swapmove_2x, a0, b0, a1, b1, n, mask, t0, t1
|
|
ushr \t0\().2d, \b0\().2d, #\n
|
|
ushr \t1\().2d, \b1\().2d, #\n
|
|
eor \t0\().16b, \t0\().16b, \a0\().16b
|
|
eor \t1\().16b, \t1\().16b, \a1\().16b
|
|
and \t0\().16b, \t0\().16b, \mask\().16b
|
|
and \t1\().16b, \t1\().16b, \mask\().16b
|
|
eor \a0\().16b, \a0\().16b, \t0\().16b
|
|
shl \t0\().2d, \t0\().2d, #\n
|
|
eor \a1\().16b, \a1\().16b, \t1\().16b
|
|
shl \t1\().2d, \t1\().2d, #\n
|
|
eor \b0\().16b, \b0\().16b, \t0\().16b
|
|
eor \b1\().16b, \b1\().16b, \t1\().16b
|
|
.endm
|
|
|
|
.macro bitslice, x7, x6, x5, x4, x3, x2, x1, x0, t0, t1, t2, t3
|
|
movi \t0\().16b, #0x55
|
|
movi \t1\().16b, #0x33
|
|
swapmove_2x \x0, \x1, \x2, \x3, 1, \t0, \t2, \t3
|
|
swapmove_2x \x4, \x5, \x6, \x7, 1, \t0, \t2, \t3
|
|
movi \t0\().16b, #0x0f
|
|
swapmove_2x \x0, \x2, \x1, \x3, 2, \t1, \t2, \t3
|
|
swapmove_2x \x4, \x6, \x5, \x7, 2, \t1, \t2, \t3
|
|
swapmove_2x \x0, \x4, \x1, \x5, 4, \t0, \t2, \t3
|
|
swapmove_2x \x2, \x6, \x3, \x7, 4, \t0, \t2, \t3
|
|
.endm
|
|
|
|
|
|
.align 6
|
|
M0: .octa 0x0004080c0105090d02060a0e03070b0f
|
|
|
|
M0SR: .octa 0x0004080c05090d010a0e02060f03070b
|
|
SR: .octa 0x0f0e0d0c0a09080b0504070600030201
|
|
SRM0: .octa 0x01060b0c0207080d0304090e00050a0f
|
|
|
|
M0ISR: .octa 0x0004080c0d0105090a0e0206070b0f03
|
|
ISR: .octa 0x0f0e0d0c080b0a090504070602010003
|
|
ISRM0: .octa 0x0306090c00070a0d01040b0e0205080f
|
|
|
|
/*
|
|
* void aesbs_convert_key(u8 out[], u32 const rk[], int rounds)
|
|
*/
|
|
SYM_FUNC_START(aesbs_convert_key)
|
|
ld1 {v7.4s}, [x1], #16 // load round 0 key
|
|
ld1 {v17.4s}, [x1], #16 // load round 1 key
|
|
|
|
movi v8.16b, #0x01 // bit masks
|
|
movi v9.16b, #0x02
|
|
movi v10.16b, #0x04
|
|
movi v11.16b, #0x08
|
|
movi v12.16b, #0x10
|
|
movi v13.16b, #0x20
|
|
movi v14.16b, #0x40
|
|
movi v15.16b, #0x80
|
|
ldr q16, M0
|
|
|
|
sub x2, x2, #1
|
|
str q7, [x0], #16 // save round 0 key
|
|
|
|
.Lkey_loop:
|
|
tbl v7.16b ,{v17.16b}, v16.16b
|
|
ld1 {v17.4s}, [x1], #16 // load next round key
|
|
|
|
cmtst v0.16b, v7.16b, v8.16b
|
|
cmtst v1.16b, v7.16b, v9.16b
|
|
cmtst v2.16b, v7.16b, v10.16b
|
|
cmtst v3.16b, v7.16b, v11.16b
|
|
cmtst v4.16b, v7.16b, v12.16b
|
|
cmtst v5.16b, v7.16b, v13.16b
|
|
cmtst v6.16b, v7.16b, v14.16b
|
|
cmtst v7.16b, v7.16b, v15.16b
|
|
not v0.16b, v0.16b
|
|
not v1.16b, v1.16b
|
|
not v5.16b, v5.16b
|
|
not v6.16b, v6.16b
|
|
|
|
subs x2, x2, #1
|
|
stp q0, q1, [x0], #128
|
|
stp q2, q3, [x0, #-96]
|
|
stp q4, q5, [x0, #-64]
|
|
stp q6, q7, [x0, #-32]
|
|
b.ne .Lkey_loop
|
|
|
|
movi v7.16b, #0x63 // compose .L63
|
|
eor v17.16b, v17.16b, v7.16b
|
|
str q17, [x0]
|
|
ret
|
|
SYM_FUNC_END(aesbs_convert_key)
|
|
|
|
.align 4
|
|
SYM_FUNC_START_LOCAL(aesbs_encrypt8)
|
|
ldr q9, [bskey], #16 // round 0 key
|
|
ldr q8, M0SR
|
|
ldr q24, SR
|
|
|
|
eor v10.16b, v0.16b, v9.16b // xor with round0 key
|
|
eor v11.16b, v1.16b, v9.16b
|
|
tbl v0.16b, {v10.16b}, v8.16b
|
|
eor v12.16b, v2.16b, v9.16b
|
|
tbl v1.16b, {v11.16b}, v8.16b
|
|
eor v13.16b, v3.16b, v9.16b
|
|
tbl v2.16b, {v12.16b}, v8.16b
|
|
eor v14.16b, v4.16b, v9.16b
|
|
tbl v3.16b, {v13.16b}, v8.16b
|
|
eor v15.16b, v5.16b, v9.16b
|
|
tbl v4.16b, {v14.16b}, v8.16b
|
|
eor v10.16b, v6.16b, v9.16b
|
|
tbl v5.16b, {v15.16b}, v8.16b
|
|
eor v11.16b, v7.16b, v9.16b
|
|
tbl v6.16b, {v10.16b}, v8.16b
|
|
tbl v7.16b, {v11.16b}, v8.16b
|
|
|
|
bitslice v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11
|
|
|
|
sub rounds, rounds, #1
|
|
b .Lenc_sbox
|
|
|
|
.Lenc_loop:
|
|
shift_rows v0, v1, v2, v3, v4, v5, v6, v7, v24
|
|
.Lenc_sbox:
|
|
sbox v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
subs rounds, rounds, #1
|
|
b.cc .Lenc_done
|
|
|
|
enc_next_rk
|
|
|
|
mix_cols v0, v1, v4, v6, v3, v7, v2, v5, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
|
|
add_round_key v0, v1, v2, v3, v4, v5, v6, v7
|
|
|
|
b.ne .Lenc_loop
|
|
ldr q24, SRM0
|
|
b .Lenc_loop
|
|
|
|
.Lenc_done:
|
|
ldr q12, [bskey] // last round key
|
|
|
|
bitslice v0, v1, v4, v6, v3, v7, v2, v5, v8, v9, v10, v11
|
|
|
|
eor v0.16b, v0.16b, v12.16b
|
|
eor v1.16b, v1.16b, v12.16b
|
|
eor v4.16b, v4.16b, v12.16b
|
|
eor v6.16b, v6.16b, v12.16b
|
|
eor v3.16b, v3.16b, v12.16b
|
|
eor v7.16b, v7.16b, v12.16b
|
|
eor v2.16b, v2.16b, v12.16b
|
|
eor v5.16b, v5.16b, v12.16b
|
|
ret
|
|
SYM_FUNC_END(aesbs_encrypt8)
|
|
|
|
.align 4
|
|
SYM_FUNC_START_LOCAL(aesbs_decrypt8)
|
|
lsl x9, rounds, #7
|
|
add bskey, bskey, x9
|
|
|
|
ldr q9, [bskey, #-112]! // round 0 key
|
|
ldr q8, M0ISR
|
|
ldr q24, ISR
|
|
|
|
eor v10.16b, v0.16b, v9.16b // xor with round0 key
|
|
eor v11.16b, v1.16b, v9.16b
|
|
tbl v0.16b, {v10.16b}, v8.16b
|
|
eor v12.16b, v2.16b, v9.16b
|
|
tbl v1.16b, {v11.16b}, v8.16b
|
|
eor v13.16b, v3.16b, v9.16b
|
|
tbl v2.16b, {v12.16b}, v8.16b
|
|
eor v14.16b, v4.16b, v9.16b
|
|
tbl v3.16b, {v13.16b}, v8.16b
|
|
eor v15.16b, v5.16b, v9.16b
|
|
tbl v4.16b, {v14.16b}, v8.16b
|
|
eor v10.16b, v6.16b, v9.16b
|
|
tbl v5.16b, {v15.16b}, v8.16b
|
|
eor v11.16b, v7.16b, v9.16b
|
|
tbl v6.16b, {v10.16b}, v8.16b
|
|
tbl v7.16b, {v11.16b}, v8.16b
|
|
|
|
bitslice v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11
|
|
|
|
sub rounds, rounds, #1
|
|
b .Ldec_sbox
|
|
|
|
.Ldec_loop:
|
|
shift_rows v0, v1, v2, v3, v4, v5, v6, v7, v24
|
|
.Ldec_sbox:
|
|
inv_sbox v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
subs rounds, rounds, #1
|
|
b.cc .Ldec_done
|
|
|
|
dec_next_rk
|
|
|
|
add_round_key v0, v1, v6, v4, v2, v7, v3, v5
|
|
|
|
inv_mix_cols v0, v1, v6, v4, v2, v7, v3, v5, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
|
|
b.ne .Ldec_loop
|
|
ldr q24, ISRM0
|
|
b .Ldec_loop
|
|
.Ldec_done:
|
|
ldr q12, [bskey, #-16] // last round key
|
|
|
|
bitslice v0, v1, v6, v4, v2, v7, v3, v5, v8, v9, v10, v11
|
|
|
|
eor v0.16b, v0.16b, v12.16b
|
|
eor v1.16b, v1.16b, v12.16b
|
|
eor v6.16b, v6.16b, v12.16b
|
|
eor v4.16b, v4.16b, v12.16b
|
|
eor v2.16b, v2.16b, v12.16b
|
|
eor v7.16b, v7.16b, v12.16b
|
|
eor v3.16b, v3.16b, v12.16b
|
|
eor v5.16b, v5.16b, v12.16b
|
|
ret
|
|
SYM_FUNC_END(aesbs_decrypt8)
|
|
|
|
/*
|
|
* aesbs_ecb_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks)
|
|
* aesbs_ecb_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks)
|
|
*/
|
|
.macro __ecb_crypt, do8, o0, o1, o2, o3, o4, o5, o6, o7
|
|
frame_push 5
|
|
|
|
mov x19, x0
|
|
mov x20, x1
|
|
mov x21, x2
|
|
mov x22, x3
|
|
mov x23, x4
|
|
|
|
99: mov x5, #1
|
|
lsl x5, x5, x23
|
|
subs w23, w23, #8
|
|
csel x23, x23, xzr, pl
|
|
csel x5, x5, xzr, mi
|
|
|
|
ld1 {v0.16b}, [x20], #16
|
|
tbnz x5, #1, 0f
|
|
ld1 {v1.16b}, [x20], #16
|
|
tbnz x5, #2, 0f
|
|
ld1 {v2.16b}, [x20], #16
|
|
tbnz x5, #3, 0f
|
|
ld1 {v3.16b}, [x20], #16
|
|
tbnz x5, #4, 0f
|
|
ld1 {v4.16b}, [x20], #16
|
|
tbnz x5, #5, 0f
|
|
ld1 {v5.16b}, [x20], #16
|
|
tbnz x5, #6, 0f
|
|
ld1 {v6.16b}, [x20], #16
|
|
tbnz x5, #7, 0f
|
|
ld1 {v7.16b}, [x20], #16
|
|
|
|
0: mov bskey, x21
|
|
mov rounds, x22
|
|
bl \do8
|
|
|
|
st1 {\o0\().16b}, [x19], #16
|
|
tbnz x5, #1, 1f
|
|
st1 {\o1\().16b}, [x19], #16
|
|
tbnz x5, #2, 1f
|
|
st1 {\o2\().16b}, [x19], #16
|
|
tbnz x5, #3, 1f
|
|
st1 {\o3\().16b}, [x19], #16
|
|
tbnz x5, #4, 1f
|
|
st1 {\o4\().16b}, [x19], #16
|
|
tbnz x5, #5, 1f
|
|
st1 {\o5\().16b}, [x19], #16
|
|
tbnz x5, #6, 1f
|
|
st1 {\o6\().16b}, [x19], #16
|
|
tbnz x5, #7, 1f
|
|
st1 {\o7\().16b}, [x19], #16
|
|
|
|
cbz x23, 1f
|
|
b 99b
|
|
|
|
1: frame_pop
|
|
ret
|
|
.endm
|
|
|
|
.align 4
|
|
SYM_TYPED_FUNC_START(aesbs_ecb_encrypt)
|
|
__ecb_crypt aesbs_encrypt8, v0, v1, v4, v6, v3, v7, v2, v5
|
|
SYM_FUNC_END(aesbs_ecb_encrypt)
|
|
|
|
.align 4
|
|
SYM_TYPED_FUNC_START(aesbs_ecb_decrypt)
|
|
__ecb_crypt aesbs_decrypt8, v0, v1, v6, v4, v2, v7, v3, v5
|
|
SYM_FUNC_END(aesbs_ecb_decrypt)
|
|
|
|
/*
|
|
* aesbs_cbc_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
*/
|
|
.align 4
|
|
SYM_FUNC_START(aesbs_cbc_decrypt)
|
|
frame_push 6
|
|
|
|
mov x19, x0
|
|
mov x20, x1
|
|
mov x21, x2
|
|
mov x22, x3
|
|
mov x23, x4
|
|
mov x24, x5
|
|
|
|
99: mov x6, #1
|
|
lsl x6, x6, x23
|
|
subs w23, w23, #8
|
|
csel x23, x23, xzr, pl
|
|
csel x6, x6, xzr, mi
|
|
|
|
ld1 {v0.16b}, [x20], #16
|
|
mov v25.16b, v0.16b
|
|
tbnz x6, #1, 0f
|
|
ld1 {v1.16b}, [x20], #16
|
|
mov v26.16b, v1.16b
|
|
tbnz x6, #2, 0f
|
|
ld1 {v2.16b}, [x20], #16
|
|
mov v27.16b, v2.16b
|
|
tbnz x6, #3, 0f
|
|
ld1 {v3.16b}, [x20], #16
|
|
mov v28.16b, v3.16b
|
|
tbnz x6, #4, 0f
|
|
ld1 {v4.16b}, [x20], #16
|
|
mov v29.16b, v4.16b
|
|
tbnz x6, #5, 0f
|
|
ld1 {v5.16b}, [x20], #16
|
|
mov v30.16b, v5.16b
|
|
tbnz x6, #6, 0f
|
|
ld1 {v6.16b}, [x20], #16
|
|
mov v31.16b, v6.16b
|
|
tbnz x6, #7, 0f
|
|
ld1 {v7.16b}, [x20]
|
|
|
|
0: mov bskey, x21
|
|
mov rounds, x22
|
|
bl aesbs_decrypt8
|
|
|
|
ld1 {v24.16b}, [x24] // load IV
|
|
|
|
eor v1.16b, v1.16b, v25.16b
|
|
eor v6.16b, v6.16b, v26.16b
|
|
eor v4.16b, v4.16b, v27.16b
|
|
eor v2.16b, v2.16b, v28.16b
|
|
eor v7.16b, v7.16b, v29.16b
|
|
eor v0.16b, v0.16b, v24.16b
|
|
eor v3.16b, v3.16b, v30.16b
|
|
eor v5.16b, v5.16b, v31.16b
|
|
|
|
st1 {v0.16b}, [x19], #16
|
|
mov v24.16b, v25.16b
|
|
tbnz x6, #1, 1f
|
|
st1 {v1.16b}, [x19], #16
|
|
mov v24.16b, v26.16b
|
|
tbnz x6, #2, 1f
|
|
st1 {v6.16b}, [x19], #16
|
|
mov v24.16b, v27.16b
|
|
tbnz x6, #3, 1f
|
|
st1 {v4.16b}, [x19], #16
|
|
mov v24.16b, v28.16b
|
|
tbnz x6, #4, 1f
|
|
st1 {v2.16b}, [x19], #16
|
|
mov v24.16b, v29.16b
|
|
tbnz x6, #5, 1f
|
|
st1 {v7.16b}, [x19], #16
|
|
mov v24.16b, v30.16b
|
|
tbnz x6, #6, 1f
|
|
st1 {v3.16b}, [x19], #16
|
|
mov v24.16b, v31.16b
|
|
tbnz x6, #7, 1f
|
|
ld1 {v24.16b}, [x20], #16
|
|
st1 {v5.16b}, [x19], #16
|
|
1: st1 {v24.16b}, [x24] // store IV
|
|
|
|
cbz x23, 2f
|
|
b 99b
|
|
|
|
2: frame_pop
|
|
ret
|
|
SYM_FUNC_END(aesbs_cbc_decrypt)
|
|
|
|
.macro next_tweak, out, in, const, tmp
|
|
sshr \tmp\().2d, \in\().2d, #63
|
|
and \tmp\().16b, \tmp\().16b, \const\().16b
|
|
add \out\().2d, \in\().2d, \in\().2d
|
|
ext \tmp\().16b, \tmp\().16b, \tmp\().16b, #8
|
|
eor \out\().16b, \out\().16b, \tmp\().16b
|
|
.endm
|
|
|
|
/*
|
|
* aesbs_xts_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
* aesbs_xts_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
*/
|
|
SYM_FUNC_START_LOCAL(__xts_crypt8)
|
|
movi v18.2s, #0x1
|
|
movi v19.2s, #0x87
|
|
uzp1 v18.4s, v18.4s, v19.4s
|
|
|
|
ld1 {v0.16b-v3.16b}, [x1], #64
|
|
ld1 {v4.16b-v7.16b}, [x1], #64
|
|
|
|
next_tweak v26, v25, v18, v19
|
|
next_tweak v27, v26, v18, v19
|
|
next_tweak v28, v27, v18, v19
|
|
next_tweak v29, v28, v18, v19
|
|
next_tweak v30, v29, v18, v19
|
|
next_tweak v31, v30, v18, v19
|
|
next_tweak v16, v31, v18, v19
|
|
next_tweak v17, v16, v18, v19
|
|
|
|
eor v0.16b, v0.16b, v25.16b
|
|
eor v1.16b, v1.16b, v26.16b
|
|
eor v2.16b, v2.16b, v27.16b
|
|
eor v3.16b, v3.16b, v28.16b
|
|
eor v4.16b, v4.16b, v29.16b
|
|
eor v5.16b, v5.16b, v30.16b
|
|
eor v6.16b, v6.16b, v31.16b
|
|
eor v7.16b, v7.16b, v16.16b
|
|
|
|
stp q16, q17, [x6]
|
|
|
|
mov bskey, x2
|
|
mov rounds, x3
|
|
br x16
|
|
SYM_FUNC_END(__xts_crypt8)
|
|
|
|
.macro __xts_crypt, do8, o0, o1, o2, o3, o4, o5, o6, o7
|
|
frame_push 0, 32
|
|
add x6, sp, #.Lframe_local_offset
|
|
|
|
ld1 {v25.16b}, [x5]
|
|
|
|
0: adr x16, \do8
|
|
bl __xts_crypt8
|
|
|
|
eor v16.16b, \o0\().16b, v25.16b
|
|
eor v17.16b, \o1\().16b, v26.16b
|
|
eor v18.16b, \o2\().16b, v27.16b
|
|
eor v19.16b, \o3\().16b, v28.16b
|
|
|
|
ldp q24, q25, [x6]
|
|
|
|
eor v20.16b, \o4\().16b, v29.16b
|
|
eor v21.16b, \o5\().16b, v30.16b
|
|
eor v22.16b, \o6\().16b, v31.16b
|
|
eor v23.16b, \o7\().16b, v24.16b
|
|
|
|
st1 {v16.16b-v19.16b}, [x0], #64
|
|
st1 {v20.16b-v23.16b}, [x0], #64
|
|
|
|
subs x4, x4, #8
|
|
b.gt 0b
|
|
|
|
st1 {v25.16b}, [x5]
|
|
frame_pop
|
|
ret
|
|
.endm
|
|
|
|
SYM_TYPED_FUNC_START(aesbs_xts_encrypt)
|
|
__xts_crypt aesbs_encrypt8, v0, v1, v4, v6, v3, v7, v2, v5
|
|
SYM_FUNC_END(aesbs_xts_encrypt)
|
|
|
|
SYM_TYPED_FUNC_START(aesbs_xts_decrypt)
|
|
__xts_crypt aesbs_decrypt8, v0, v1, v6, v4, v2, v7, v3, v5
|
|
SYM_FUNC_END(aesbs_xts_decrypt)
|
|
|
|
.macro next_ctr, v
|
|
mov \v\().d[1], x8
|
|
adds x8, x8, #1
|
|
mov \v\().d[0], x7
|
|
adc x7, x7, xzr
|
|
rev64 \v\().16b, \v\().16b
|
|
.endm
|
|
|
|
/*
|
|
* aesbs_ctr_encrypt(u8 out[], u8 const in[], u8 const rk[],
|
|
* int rounds, int blocks, u8 iv[])
|
|
*/
|
|
SYM_FUNC_START(aesbs_ctr_encrypt)
|
|
frame_push 0
|
|
ldp x7, x8, [x5]
|
|
ld1 {v0.16b}, [x5]
|
|
CPU_LE( rev x7, x7 )
|
|
CPU_LE( rev x8, x8 )
|
|
adds x8, x8, #1
|
|
adc x7, x7, xzr
|
|
|
|
0: next_ctr v1
|
|
next_ctr v2
|
|
next_ctr v3
|
|
next_ctr v4
|
|
next_ctr v5
|
|
next_ctr v6
|
|
next_ctr v7
|
|
|
|
mov bskey, x2
|
|
mov rounds, x3
|
|
bl aesbs_encrypt8
|
|
|
|
ld1 { v8.16b-v11.16b}, [x1], #64
|
|
ld1 {v12.16b-v15.16b}, [x1], #64
|
|
|
|
eor v8.16b, v0.16b, v8.16b
|
|
eor v9.16b, v1.16b, v9.16b
|
|
eor v10.16b, v4.16b, v10.16b
|
|
eor v11.16b, v6.16b, v11.16b
|
|
eor v12.16b, v3.16b, v12.16b
|
|
eor v13.16b, v7.16b, v13.16b
|
|
eor v14.16b, v2.16b, v14.16b
|
|
eor v15.16b, v5.16b, v15.16b
|
|
|
|
st1 { v8.16b-v11.16b}, [x0], #64
|
|
st1 {v12.16b-v15.16b}, [x0], #64
|
|
|
|
next_ctr v0
|
|
subs x4, x4, #8
|
|
b.gt 0b
|
|
|
|
st1 {v0.16b}, [x5]
|
|
frame_pop
|
|
ret
|
|
SYM_FUNC_END(aesbs_ctr_encrypt)
|