mirror of
https://github.com/torvalds/linux.git
synced 2024-12-20 10:01:56 +00:00
187818cd6a
Document what's going on in asm/backoff.h with a large and descriptive comment. Refer to it above the cpu_relax() definition in asm/processor_64.h Rename the pause patching section to have "3insn" in it's name like the other patching sections do. Based upon feedback from Sam Ravnborg. Signed-off-by: David S. Miller <davem@davemloft.net>
86 lines
2.7 KiB
C
86 lines
2.7 KiB
C
#ifndef _SPARC64_BACKOFF_H
|
|
#define _SPARC64_BACKOFF_H
|
|
|
|
/* The macros in this file implement an exponential backoff facility
|
|
* for atomic operations.
|
|
*
|
|
* When multiple threads compete on an atomic operation, it is
|
|
* possible for one thread to be continually denied a successful
|
|
* completion of the compare-and-swap instruction. Heavily
|
|
* threaded cpu implementations like Niagara can compound this
|
|
* problem even further.
|
|
*
|
|
* When an atomic operation fails and needs to be retried, we spin a
|
|
* certain number of times. At each subsequent failure of the same
|
|
* operation we double the spin count, realizing an exponential
|
|
* backoff.
|
|
*
|
|
* When we spin, we try to use an operation that will cause the
|
|
* current cpu strand to block, and therefore make the core fully
|
|
* available to any other other runnable strands. There are two
|
|
* options, based upon cpu capabilities.
|
|
*
|
|
* On all cpus prior to SPARC-T4 we do three dummy reads of the
|
|
* condition code register. Each read blocks the strand for something
|
|
* between 40 and 50 cpu cycles.
|
|
*
|
|
* For SPARC-T4 and later we have a special "pause" instruction
|
|
* available. This is implemented using writes to register %asr27.
|
|
* The cpu will block the number of cycles written into the register,
|
|
* unless a disrupting trap happens first. SPARC-T4 specifically
|
|
* implements pause with a granularity of 8 cycles. Each strand has
|
|
* an internal pause counter which decrements every 8 cycles. So the
|
|
* chip shifts the %asr27 value down by 3 bits, and writes the result
|
|
* into the pause counter. If a value smaller than 8 is written, the
|
|
* chip blocks for 1 cycle.
|
|
*
|
|
* To achieve the same amount of backoff as the three %ccr reads give
|
|
* on earlier chips, we shift the backoff value up by 7 bits. (Three
|
|
* %ccr reads block for about 128 cycles, 1 << 7 == 128) We write the
|
|
* whole amount we want to block into the pause register, rather than
|
|
* loop writing 128 each time.
|
|
*/
|
|
|
|
#define BACKOFF_LIMIT (4 * 1024)
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
#define BACKOFF_SETUP(reg) \
|
|
mov 1, reg
|
|
|
|
#define BACKOFF_LABEL(spin_label, continue_label) \
|
|
spin_label
|
|
|
|
#define BACKOFF_SPIN(reg, tmp, label) \
|
|
mov reg, tmp; \
|
|
88: rd %ccr, %g0; \
|
|
rd %ccr, %g0; \
|
|
rd %ccr, %g0; \
|
|
.section .pause_3insn_patch,"ax";\
|
|
.word 88b; \
|
|
sllx tmp, 7, tmp; \
|
|
wr tmp, 0, %asr27; \
|
|
clr tmp; \
|
|
.previous; \
|
|
brnz,pt tmp, 88b; \
|
|
sub tmp, 1, tmp; \
|
|
set BACKOFF_LIMIT, tmp; \
|
|
cmp reg, tmp; \
|
|
bg,pn %xcc, label; \
|
|
nop; \
|
|
ba,pt %xcc, label; \
|
|
sllx reg, 1, reg;
|
|
|
|
#else
|
|
|
|
#define BACKOFF_SETUP(reg)
|
|
|
|
#define BACKOFF_LABEL(spin_label, continue_label) \
|
|
continue_label
|
|
|
|
#define BACKOFF_SPIN(reg, tmp, label)
|
|
|
|
#endif
|
|
|
|
#endif /* _SPARC64_BACKOFF_H */
|