mirror of
https://github.com/torvalds/linux.git
synced 2024-12-02 00:51:44 +00:00
f94909ceb1
Replace all ret/retq instructions with RET in preparation of making RET a macro. Since AS is case insensitive it's a big no-op without RET defined. find arch/x86/ -name \*.S | while read file do sed -i 's/\<ret[q]*\>/RET/' $file done Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Signed-off-by: Borislav Petkov <bp@suse.de> Link: https://lore.kernel.org/r/20211204134907.905503893@infradead.org
180 lines
4.2 KiB
ArmAsm
180 lines
4.2 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*---------------------------------------------------------------------------+
|
|
| mul_Xsig.S |
|
|
| |
|
|
| Multiply a 12 byte fixed point number by another fixed point number. |
|
|
| |
|
|
| Copyright (C) 1992,1994,1995 |
|
|
| W. Metzenthen, 22 Parker St, Ormond, Vic 3163, |
|
|
| Australia. E-mail billm@jacobi.maths.monash.edu.au |
|
|
| |
|
|
| Call from C as: |
|
|
| void mul32_Xsig(Xsig *x, unsigned b) |
|
|
| |
|
|
| void mul64_Xsig(Xsig *x, unsigned long long *b) |
|
|
| |
|
|
| void mul_Xsig_Xsig(Xsig *x, unsigned *b) |
|
|
| |
|
|
| The result is neither rounded nor normalized, and the ls bit or so may |
|
|
| be wrong. |
|
|
| |
|
|
+---------------------------------------------------------------------------*/
|
|
.file "mul_Xsig.S"
|
|
|
|
|
|
#include "fpu_emu.h"
|
|
|
|
.text
|
|
SYM_FUNC_START(mul32_Xsig)
|
|
pushl %ebp
|
|
movl %esp,%ebp
|
|
subl $16,%esp
|
|
pushl %esi
|
|
|
|
movl PARAM1,%esi
|
|
movl PARAM2,%ecx
|
|
|
|
xor %eax,%eax
|
|
movl %eax,-4(%ebp)
|
|
movl %eax,-8(%ebp)
|
|
|
|
movl (%esi),%eax /* lsl of Xsig */
|
|
mull %ecx /* msl of b */
|
|
movl %edx,-12(%ebp)
|
|
|
|
movl 4(%esi),%eax /* midl of Xsig */
|
|
mull %ecx /* msl of b */
|
|
addl %eax,-12(%ebp)
|
|
adcl %edx,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 8(%esi),%eax /* msl of Xsig */
|
|
mull %ecx /* msl of b */
|
|
addl %eax,-8(%ebp)
|
|
adcl %edx,-4(%ebp)
|
|
|
|
movl -12(%ebp),%eax
|
|
movl %eax,(%esi)
|
|
movl -8(%ebp),%eax
|
|
movl %eax,4(%esi)
|
|
movl -4(%ebp),%eax
|
|
movl %eax,8(%esi)
|
|
|
|
popl %esi
|
|
leave
|
|
RET
|
|
SYM_FUNC_END(mul32_Xsig)
|
|
|
|
|
|
SYM_FUNC_START(mul64_Xsig)
|
|
pushl %ebp
|
|
movl %esp,%ebp
|
|
subl $16,%esp
|
|
pushl %esi
|
|
|
|
movl PARAM1,%esi
|
|
movl PARAM2,%ecx
|
|
|
|
xor %eax,%eax
|
|
movl %eax,-4(%ebp)
|
|
movl %eax,-8(%ebp)
|
|
|
|
movl (%esi),%eax /* lsl of Xsig */
|
|
mull 4(%ecx) /* msl of b */
|
|
movl %edx,-12(%ebp)
|
|
|
|
movl 4(%esi),%eax /* midl of Xsig */
|
|
mull (%ecx) /* lsl of b */
|
|
addl %edx,-12(%ebp)
|
|
adcl $0,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 4(%esi),%eax /* midl of Xsig */
|
|
mull 4(%ecx) /* msl of b */
|
|
addl %eax,-12(%ebp)
|
|
adcl %edx,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 8(%esi),%eax /* msl of Xsig */
|
|
mull (%ecx) /* lsl of b */
|
|
addl %eax,-12(%ebp)
|
|
adcl %edx,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 8(%esi),%eax /* msl of Xsig */
|
|
mull 4(%ecx) /* msl of b */
|
|
addl %eax,-8(%ebp)
|
|
adcl %edx,-4(%ebp)
|
|
|
|
movl -12(%ebp),%eax
|
|
movl %eax,(%esi)
|
|
movl -8(%ebp),%eax
|
|
movl %eax,4(%esi)
|
|
movl -4(%ebp),%eax
|
|
movl %eax,8(%esi)
|
|
|
|
popl %esi
|
|
leave
|
|
RET
|
|
SYM_FUNC_END(mul64_Xsig)
|
|
|
|
|
|
|
|
SYM_FUNC_START(mul_Xsig_Xsig)
|
|
pushl %ebp
|
|
movl %esp,%ebp
|
|
subl $16,%esp
|
|
pushl %esi
|
|
|
|
movl PARAM1,%esi
|
|
movl PARAM2,%ecx
|
|
|
|
xor %eax,%eax
|
|
movl %eax,-4(%ebp)
|
|
movl %eax,-8(%ebp)
|
|
|
|
movl (%esi),%eax /* lsl of Xsig */
|
|
mull 8(%ecx) /* msl of b */
|
|
movl %edx,-12(%ebp)
|
|
|
|
movl 4(%esi),%eax /* midl of Xsig */
|
|
mull 4(%ecx) /* midl of b */
|
|
addl %edx,-12(%ebp)
|
|
adcl $0,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 8(%esi),%eax /* msl of Xsig */
|
|
mull (%ecx) /* lsl of b */
|
|
addl %edx,-12(%ebp)
|
|
adcl $0,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 4(%esi),%eax /* midl of Xsig */
|
|
mull 8(%ecx) /* msl of b */
|
|
addl %eax,-12(%ebp)
|
|
adcl %edx,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 8(%esi),%eax /* msl of Xsig */
|
|
mull 4(%ecx) /* midl of b */
|
|
addl %eax,-12(%ebp)
|
|
adcl %edx,-8(%ebp)
|
|
adcl $0,-4(%ebp)
|
|
|
|
movl 8(%esi),%eax /* msl of Xsig */
|
|
mull 8(%ecx) /* msl of b */
|
|
addl %eax,-8(%ebp)
|
|
adcl %edx,-4(%ebp)
|
|
|
|
movl -12(%ebp),%edx
|
|
movl %edx,(%esi)
|
|
movl -8(%ebp),%edx
|
|
movl %edx,4(%esi)
|
|
movl -4(%ebp),%edx
|
|
movl %edx,8(%esi)
|
|
|
|
popl %esi
|
|
leave
|
|
RET
|
|
SYM_FUNC_END(mul_Xsig_Xsig)
|