forked from Minki/linux
df9ee29270
Fix the IRQ flag handling naming. In linux/irqflags.h under one configuration, it maps: local_irq_enable() -> raw_local_irq_enable() local_irq_disable() -> raw_local_irq_disable() local_irq_save() -> raw_local_irq_save() ... and under the other configuration, it maps: raw_local_irq_enable() -> local_irq_enable() raw_local_irq_disable() -> local_irq_disable() raw_local_irq_save() -> local_irq_save() ... This is quite confusing. There should be one set of names expected of the arch, and this should be wrapped to give another set of names that are expected by users of this facility. Change this to have the arch provide: flags = arch_local_save_flags() flags = arch_local_irq_save() arch_local_irq_restore(flags) arch_local_irq_disable() arch_local_irq_enable() arch_irqs_disabled_flags(flags) arch_irqs_disabled() arch_safe_halt() Then linux/irqflags.h wraps these to provide: raw_local_save_flags(flags) raw_local_irq_save(flags) raw_local_irq_restore(flags) raw_local_irq_disable() raw_local_irq_enable() raw_irqs_disabled_flags(flags) raw_irqs_disabled() raw_safe_halt() with type checking on the flags 'arguments', and then wraps those to provide: local_save_flags(flags) local_irq_save(flags) local_irq_restore(flags) local_irq_disable() local_irq_enable() irqs_disabled_flags(flags) irqs_disabled() safe_halt() with tracing included if enabled. The arch functions can now all be inline functions rather than some of them having to be macros. Signed-off-by: David Howells <dhowells@redhat.com> [X86, FRV, MN10300] Signed-off-by: Chris Metcalf <cmetcalf@tilera.com> [Tile] Signed-off-by: Michal Simek <monstr@monstr.eu> [Microblaze] Tested-by: Catalin Marinas <catalin.marinas@arm.com> [ARM] Acked-by: Thomas Gleixner <tglx@linutronix.de> Acked-by: Haavard Skinnemoen <haavard.skinnemoen@atmel.com> [AVR] Acked-by: Tony Luck <tony.luck@intel.com> [IA-64] Acked-by: Hirokazu Takata <takata@linux-m32r.org> [M32R] Acked-by: Greg Ungerer <gerg@uclinux.org> [M68K/M68KNOMMU] Acked-by: Ralf Baechle <ralf@linux-mips.org> [MIPS] Acked-by: Kyle McMartin <kyle@mcmartin.ca> [PA-RISC] Acked-by: Paul Mackerras <paulus@samba.org> [PowerPC] Acked-by: Martin Schwidefsky <schwidefsky@de.ibm.com> [S390] Acked-by: Chen Liqin <liqin.chen@sunplusct.com> [Score] Acked-by: Matt Fleming <matt@console-pimps.org> [SH] Acked-by: David S. Miller <davem@davemloft.net> [Sparc] Acked-by: Chris Zankel <chris@zankel.net> [Xtensa] Reviewed-by: Richard Henderson <rth@twiddle.net> [Alpha] Reviewed-by: Yoshinori Sato <ysato@users.sourceforge.jp> [H8300] Cc: starvik@axis.com [CRIS] Cc: jesper.nilsson@axis.com [CRIS] Cc: linux-cris-kernel@axis.com
185 lines
4.8 KiB
C
185 lines
4.8 KiB
C
/*
|
|
* include/asm-xtensa/system.h
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (C) 2001 - 2005 Tensilica Inc.
|
|
*/
|
|
|
|
#ifndef _XTENSA_SYSTEM_H
|
|
#define _XTENSA_SYSTEM_H
|
|
|
|
#include <linux/stringify.h>
|
|
#include <linux/irqflags.h>
|
|
|
|
#include <asm/processor.h>
|
|
|
|
#define smp_read_barrier_depends() do { } while(0)
|
|
#define read_barrier_depends() do { } while(0)
|
|
|
|
#define mb() barrier()
|
|
#define rmb() mb()
|
|
#define wmb() mb()
|
|
|
|
#ifdef CONFIG_SMP
|
|
#error smp_* not defined
|
|
#else
|
|
#define smp_mb() barrier()
|
|
#define smp_rmb() barrier()
|
|
#define smp_wmb() barrier()
|
|
#endif
|
|
|
|
#define set_mb(var, value) do { var = value; mb(); } while (0)
|
|
|
|
#if !defined (__ASSEMBLY__)
|
|
|
|
/* * switch_to(n) should switch tasks to task nr n, first
|
|
* checking that n isn't the current task, in which case it does nothing.
|
|
*/
|
|
extern void *_switch_to(void *last, void *next);
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#define switch_to(prev,next,last) \
|
|
do { \
|
|
(last) = _switch_to(prev, next); \
|
|
} while(0)
|
|
|
|
/*
|
|
* cmpxchg
|
|
*/
|
|
|
|
static inline unsigned long
|
|
__cmpxchg_u32(volatile int *p, int old, int new)
|
|
{
|
|
__asm__ __volatile__("rsil a15, "__stringify(LOCKLEVEL)"\n\t"
|
|
"l32i %0, %1, 0 \n\t"
|
|
"bne %0, %2, 1f \n\t"
|
|
"s32i %3, %1, 0 \n\t"
|
|
"1: \n\t"
|
|
"wsr a15, "__stringify(PS)" \n\t"
|
|
"rsync \n\t"
|
|
: "=&a" (old)
|
|
: "a" (p), "a" (old), "r" (new)
|
|
: "a15", "memory");
|
|
return old;
|
|
}
|
|
/* This function doesn't exist, so you'll get a linker error
|
|
* if something tries to do an invalid cmpxchg(). */
|
|
|
|
extern void __cmpxchg_called_with_bad_pointer(void);
|
|
|
|
static __inline__ unsigned long
|
|
__cmpxchg(volatile void *ptr, unsigned long old, unsigned long new, int size)
|
|
{
|
|
switch (size) {
|
|
case 4: return __cmpxchg_u32(ptr, old, new);
|
|
default: __cmpxchg_called_with_bad_pointer();
|
|
return old;
|
|
}
|
|
}
|
|
|
|
#define cmpxchg(ptr,o,n) \
|
|
({ __typeof__(*(ptr)) _o_ = (o); \
|
|
__typeof__(*(ptr)) _n_ = (n); \
|
|
(__typeof__(*(ptr))) __cmpxchg((ptr), (unsigned long)_o_, \
|
|
(unsigned long)_n_, sizeof (*(ptr))); \
|
|
})
|
|
|
|
#include <asm-generic/cmpxchg-local.h>
|
|
|
|
static inline unsigned long __cmpxchg_local(volatile void *ptr,
|
|
unsigned long old,
|
|
unsigned long new, int size)
|
|
{
|
|
switch (size) {
|
|
case 4:
|
|
return __cmpxchg_u32(ptr, old, new);
|
|
default:
|
|
return __cmpxchg_local_generic(ptr, old, new, size);
|
|
}
|
|
|
|
return old;
|
|
}
|
|
|
|
/*
|
|
* cmpxchg_local and cmpxchg64_local are atomic wrt current CPU. Always make
|
|
* them available.
|
|
*/
|
|
#define cmpxchg_local(ptr, o, n) \
|
|
((__typeof__(*(ptr)))__cmpxchg_local_generic((ptr), (unsigned long)(o),\
|
|
(unsigned long)(n), sizeof(*(ptr))))
|
|
#define cmpxchg64_local(ptr, o, n) __cmpxchg64_local_generic((ptr), (o), (n))
|
|
|
|
/*
|
|
* xchg_u32
|
|
*
|
|
* Note that a15 is used here because the register allocation
|
|
* done by the compiler is not guaranteed and a window overflow
|
|
* may not occur between the rsil and wsr instructions. By using
|
|
* a15 in the rsil, the machine is guaranteed to be in a state
|
|
* where no register reference will cause an overflow.
|
|
*/
|
|
|
|
static inline unsigned long xchg_u32(volatile int * m, unsigned long val)
|
|
{
|
|
unsigned long tmp;
|
|
__asm__ __volatile__("rsil a15, "__stringify(LOCKLEVEL)"\n\t"
|
|
"l32i %0, %1, 0 \n\t"
|
|
"s32i %2, %1, 0 \n\t"
|
|
"wsr a15, "__stringify(PS)" \n\t"
|
|
"rsync \n\t"
|
|
: "=&a" (tmp)
|
|
: "a" (m), "a" (val)
|
|
: "a15", "memory");
|
|
return tmp;
|
|
}
|
|
|
|
#define xchg(ptr,x) ((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr))))
|
|
|
|
/*
|
|
* This only works if the compiler isn't horribly bad at optimizing.
|
|
* gcc-2.5.8 reportedly can't handle this, but I define that one to
|
|
* be dead anyway.
|
|
*/
|
|
|
|
extern void __xchg_called_with_bad_pointer(void);
|
|
|
|
static __inline__ unsigned long
|
|
__xchg(unsigned long x, volatile void * ptr, int size)
|
|
{
|
|
switch (size) {
|
|
case 4:
|
|
return xchg_u32(ptr, x);
|
|
}
|
|
__xchg_called_with_bad_pointer();
|
|
return x;
|
|
}
|
|
|
|
extern void set_except_vector(int n, void *addr);
|
|
|
|
static inline void spill_registers(void)
|
|
{
|
|
unsigned int a0, ps;
|
|
|
|
__asm__ __volatile__ (
|
|
"movi a14," __stringify (PS_EXCM_BIT) " | 1\n\t"
|
|
"mov a12, a0\n\t"
|
|
"rsr a13," __stringify(SAR) "\n\t"
|
|
"xsr a14," __stringify(PS) "\n\t"
|
|
"movi a0, _spill_registers\n\t"
|
|
"rsync\n\t"
|
|
"callx0 a0\n\t"
|
|
"mov a0, a12\n\t"
|
|
"wsr a13," __stringify(SAR) "\n\t"
|
|
"wsr a14," __stringify(PS) "\n\t"
|
|
:: "a" (&a0), "a" (&ps)
|
|
: "a2", "a3", "a4", "a7", "a11", "a12", "a13", "a14", "a15", "memory");
|
|
}
|
|
|
|
#define arch_align_stack(x) (x)
|
|
|
|
#endif /* _XTENSA_SYSTEM_H */
|