ARM: 9058/1: cache-v7: refactor v7_invalidate_l1 to avoid clobbering r5/r6
The cache invalidation code in v7_invalidate_l1 can be tweaked to re-read the associativity from CCSIDR, and keep the way identifier component in a single register that is assigned in the outer loop. This way, we need 2 registers less. Given that the number of sets is typically much larger than the associativity, rearrange the code so that the outer loop has the fewer number of iterations, ensuring that the re-read of CCSIDR only occurs a handful of times in practice. Fix the whitespace while at it, and update the comment to indicate that this code is no longer a clone of anything else. Acked-by: Nicolas Pitre <nico@fluxnic.net> Signed-off-by: Ard Biesheuvel <ardb@kernel.org> Signed-off-by: Russell King <rmk+kernel@armlinux.org.uk>
This commit is contained in:
committed by
Russell King
parent
c0e50736e8
commit
f9e7a99fb6
@@ -33,9 +33,8 @@ icache_size:
|
|||||||
* processor. We fix this by performing an invalidate, rather than a
|
* processor. We fix this by performing an invalidate, rather than a
|
||||||
* clean + invalidate, before jumping into the kernel.
|
* clean + invalidate, before jumping into the kernel.
|
||||||
*
|
*
|
||||||
* This function is cloned from arch/arm/mach-tegra/headsmp.S, and needs
|
* This function needs to be called for both secondary cores startup and
|
||||||
* to be called for both secondary cores startup and primary core resume
|
* primary core resume procedures.
|
||||||
* procedures.
|
|
||||||
*/
|
*/
|
||||||
ENTRY(v7_invalidate_l1)
|
ENTRY(v7_invalidate_l1)
|
||||||
mov r0, #0
|
mov r0, #0
|
||||||
@@ -43,32 +42,32 @@ ENTRY(v7_invalidate_l1)
|
|||||||
isb
|
isb
|
||||||
mrc p15, 1, r0, c0, c0, 0 @ read cache geometry from CCSIDR
|
mrc p15, 1, r0, c0, c0, 0 @ read cache geometry from CCSIDR
|
||||||
|
|
||||||
movw r1, #0x7fff
|
movw r3, #0x3ff
|
||||||
and r2, r1, r0, lsr #13
|
and r3, r3, r0, lsr #3 @ 'Associativity' in CCSIDR[12:3]
|
||||||
|
clz r1, r3 @ WayShift
|
||||||
|
mov r2, #1
|
||||||
|
mov r3, r3, lsl r1 @ NumWays-1 shifted into bits [31:...]
|
||||||
|
movs r1, r2, lsl r1 @ #1 shifted left by same amount
|
||||||
|
moveq r1, #1 @ r1 needs value > 0 even if only 1 way
|
||||||
|
|
||||||
movw r1, #0x3ff
|
and r2, r0, #0x7
|
||||||
|
add r2, r2, #4 @ SetShift
|
||||||
|
|
||||||
and r3, r1, r0, lsr #3 @ NumWays - 1
|
1: movw r4, #0x7fff
|
||||||
add r2, r2, #1 @ NumSets
|
and r0, r4, r0, lsr #13 @ 'NumSets' in CCSIDR[27:13]
|
||||||
|
|
||||||
and r0, r0, #0x7
|
2: mov r4, r0, lsl r2 @ NumSet << SetShift
|
||||||
add r0, r0, #4 @ SetShift
|
orr r4, r4, r3 @ Reg = (Temp<<WayShift)|(NumSets<<SetShift)
|
||||||
|
mcr p15, 0, r4, c7, c6, 2
|
||||||
clz r1, r3 @ WayShift
|
subs r0, r0, #1 @ Set--
|
||||||
add r4, r3, #1 @ NumWays
|
bpl 2b
|
||||||
1: sub r2, r2, #1 @ NumSets--
|
subs r3, r3, r1 @ Way--
|
||||||
mov r3, r4 @ Temp = NumWays
|
bcc 3f
|
||||||
2: subs r3, r3, #1 @ Temp--
|
mrc p15, 1, r0, c0, c0, 0 @ re-read cache geometry from CCSIDR
|
||||||
mov r5, r3, lsl r1
|
b 1b
|
||||||
mov r6, r2, lsl r0
|
3: dsb st
|
||||||
orr r5, r5, r6 @ Reg = (Temp<<WayShift)|(NumSets<<SetShift)
|
isb
|
||||||
mcr p15, 0, r5, c7, c6, 2
|
ret lr
|
||||||
bgt 2b
|
|
||||||
cmp r2, #0
|
|
||||||
bgt 1b
|
|
||||||
dsb st
|
|
||||||
isb
|
|
||||||
ret lr
|
|
||||||
ENDPROC(v7_invalidate_l1)
|
ENDPROC(v7_invalidate_l1)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|||||||
Reference in New Issue
Block a user