mirror of
https://github.com/torvalds/linux.git
synced 2024-12-25 12:21:37 +00:00
18269c0fd4
Identical named files renamed to <name>_32.S Refactored Makefile to prepare for unification. Linking order was altered slightly - but this is a lib.a file so it should not matter. Signed-off-by: Sam Ravnborg <sam@ravnborg.org> Signed-off-by: David S. Miller <davem@davemloft.net>
100 lines
2.6 KiB
ArmAsm
100 lines
2.6 KiB
ArmAsm
/* atomic.S: Move this stuff here for better ICACHE hit rates.
|
|
*
|
|
* Copyright (C) 1996 David S. Miller (davem@caipfs.rutgers.edu)
|
|
*/
|
|
|
|
#include <asm/ptrace.h>
|
|
#include <asm/psr.h>
|
|
|
|
.text
|
|
.align 4
|
|
|
|
.globl __atomic_begin
|
|
__atomic_begin:
|
|
|
|
#ifndef CONFIG_SMP
|
|
.globl ___xchg32_sun4c
|
|
___xchg32_sun4c:
|
|
rd %psr, %g3
|
|
andcc %g3, PSR_PIL, %g0
|
|
bne 1f
|
|
nop
|
|
wr %g3, PSR_PIL, %psr
|
|
nop; nop; nop
|
|
1:
|
|
andcc %g3, PSR_PIL, %g0
|
|
ld [%g1], %g7
|
|
bne 1f
|
|
st %g2, [%g1]
|
|
wr %g3, 0x0, %psr
|
|
nop; nop; nop
|
|
1:
|
|
mov %g7, %g2
|
|
jmpl %o7 + 8, %g0
|
|
mov %g4, %o7
|
|
|
|
.globl ___xchg32_sun4md
|
|
___xchg32_sun4md:
|
|
swap [%g1], %g2
|
|
jmpl %o7 + 8, %g0
|
|
mov %g4, %o7
|
|
#endif
|
|
|
|
/* Read asm-sparc/atomic.h carefully to understand how this works for SMP.
|
|
* Really, some things here for SMP are overly clever, go read the header.
|
|
*/
|
|
.globl ___atomic24_add
|
|
___atomic24_add:
|
|
rd %psr, %g3 ! Keep the code small, old way was stupid
|
|
nop; nop; nop; ! Let the bits set
|
|
or %g3, PSR_PIL, %g7 ! Disable interrupts
|
|
wr %g7, 0x0, %psr ! Set %psr
|
|
nop; nop; nop; ! Let the bits set
|
|
#ifdef CONFIG_SMP
|
|
1: ldstub [%g1 + 3], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 1b ! Nope...
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
sra %g7, 8, %g7 ! Get signed 24-bit integer
|
|
add %g7, %g2, %g2 ! Add in argument
|
|
sll %g2, 8, %g7 ! Transpose back to atomic24_t
|
|
st %g7, [%g1] ! Clever: This releases the lock as well.
|
|
#else
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
add %g7, %g2, %g2 ! Add in argument
|
|
st %g2, [%g1] ! Store it back
|
|
#endif
|
|
wr %g3, 0x0, %psr ! Restore original PSR_PIL
|
|
nop; nop; nop; ! Let the bits set
|
|
jmpl %o7, %g0 ! NOTE: not + 8, see callers in atomic.h
|
|
mov %g4, %o7 ! Restore %o7
|
|
|
|
.globl ___atomic24_sub
|
|
___atomic24_sub:
|
|
rd %psr, %g3 ! Keep the code small, old way was stupid
|
|
nop; nop; nop; ! Let the bits set
|
|
or %g3, PSR_PIL, %g7 ! Disable interrupts
|
|
wr %g7, 0x0, %psr ! Set %psr
|
|
nop; nop; nop; ! Let the bits set
|
|
#ifdef CONFIG_SMP
|
|
1: ldstub [%g1 + 3], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 1b ! Nope...
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
sra %g7, 8, %g7 ! Get signed 24-bit integer
|
|
sub %g7, %g2, %g2 ! Subtract argument
|
|
sll %g2, 8, %g7 ! Transpose back to atomic24_t
|
|
st %g7, [%g1] ! Clever: This releases the lock as well
|
|
#else
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
sub %g7, %g2, %g2 ! Subtract argument
|
|
st %g2, [%g1] ! Store it back
|
|
#endif
|
|
wr %g3, 0x0, %psr ! Restore original PSR_PIL
|
|
nop; nop; nop; ! Let the bits set
|
|
jmpl %o7, %g0 ! NOTE: not + 8, see callers in atomic.h
|
|
mov %g4, %o7 ! Restore %o7
|
|
|
|
.globl __atomic_end
|
|
__atomic_end:
|