forked from Minki/linux
5280e004fc
- Special consideration for IA64: Add the ability to specify arch specific per cpu flags - remove .data.percpu attribute from DEFINE_PER_CPU for non-smp case. The arch definitions are all the same. So move them into linux/percpu.h. We cannot move DECLARE_PER_CPU since some include files just include asm/percpu.h to avoid include recursion problems. Cc: Rusty Russell <rusty@rustcorp.com.au> Cc: Andi Kleen <ak@suse.de> Signed-off-by: Christoph Lameter <clameter@sgi.com> Signed-off-by: Mike Travis <travis@sgi.com> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Ingo Molnar <mingo@elte.hu>
62 lines
1.8 KiB
C
62 lines
1.8 KiB
C
#ifndef __ARCH_S390_PERCPU__
|
|
#define __ARCH_S390_PERCPU__
|
|
|
|
#include <linux/compiler.h>
|
|
#include <asm/lowcore.h>
|
|
|
|
/*
|
|
* s390 uses its own implementation for per cpu data, the offset of
|
|
* the cpu local data area is cached in the cpu's lowcore memory.
|
|
* For 64 bit module code s390 forces the use of a GOT slot for the
|
|
* address of the per cpu variable. This is needed because the module
|
|
* may be more than 4G above the per cpu area.
|
|
*/
|
|
#if defined(__s390x__) && defined(MODULE)
|
|
|
|
#define __reloc_hide(var,offset) (*({ \
|
|
extern int simple_identifier_##var(void); \
|
|
unsigned long *__ptr; \
|
|
asm ( "larl %0,per_cpu__"#var"@GOTENT" \
|
|
: "=a" (__ptr) : "X" (per_cpu__##var) ); \
|
|
(typeof(&per_cpu__##var))((*__ptr) + (offset)); }))
|
|
|
|
#else
|
|
|
|
#define __reloc_hide(var, offset) (*({ \
|
|
extern int simple_identifier_##var(void); \
|
|
unsigned long __ptr; \
|
|
asm ( "" : "=a" (__ptr) : "0" (&per_cpu__##var) ); \
|
|
(typeof(&per_cpu__##var)) (__ptr + (offset)); }))
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
extern unsigned long __per_cpu_offset[NR_CPUS];
|
|
|
|
#define __get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset)
|
|
#define __raw_get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset)
|
|
#define per_cpu(var,cpu) __reloc_hide(var,__per_cpu_offset[cpu])
|
|
#define per_cpu_offset(x) (__per_cpu_offset[x])
|
|
|
|
/* A macro to avoid #include hell... */
|
|
#define percpu_modcopy(pcpudst, src, size) \
|
|
do { \
|
|
unsigned int __i; \
|
|
for_each_possible_cpu(__i) \
|
|
memcpy((pcpudst)+__per_cpu_offset[__i], \
|
|
(src), (size)); \
|
|
} while (0)
|
|
|
|
#else /* ! SMP */
|
|
|
|
#define __get_cpu_var(var) __reloc_hide(var,0)
|
|
#define __raw_get_cpu_var(var) __reloc_hide(var,0)
|
|
#define per_cpu(var,cpu) __reloc_hide(var,0)
|
|
|
|
#endif /* SMP */
|
|
|
|
#define DECLARE_PER_CPU(type, name) extern __typeof__(type) per_cpu__##name
|
|
|
|
#endif /* __ARCH_S390_PERCPU__ */
|