forked from Minki/linux
e8c10ef9dd
Change static bios_cpu_apicid array to a per_cpu data variable. This includes using a static array used during initialization similar to the way x86_cpu_to_apicid[] is handled. There is one early use of bios_cpu_apicid in apic_is_clustered_box(). The other reference in cpu_present_to_apicid() is called after smp_set_apicids() has setup the percpu version of bios_cpu_apicid. [ mingo@elte.hu: build fix ] Signed-off-by: Mike Travis <travis@sgi.com> Reviewed-by: Christoph Lameter <clameter@sgi.com> Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
102 lines
2.5 KiB
C
102 lines
2.5 KiB
C
#ifndef __ASM_SMP_H
|
|
#define __ASM_SMP_H
|
|
|
|
#include <linux/cpumask.h>
|
|
#include <linux/init.h>
|
|
|
|
/*
|
|
* We need the APIC definitions automatically as part of 'smp.h'
|
|
*/
|
|
#include <asm/apic.h>
|
|
#include <asm/io_apic.h>
|
|
#include <asm/mpspec.h>
|
|
#include <asm/pda.h>
|
|
#include <asm/thread_info.h>
|
|
|
|
extern cpumask_t cpu_callout_map;
|
|
extern cpumask_t cpu_initialized;
|
|
|
|
extern int smp_num_siblings;
|
|
extern unsigned int num_processors;
|
|
|
|
extern void smp_alloc_memory(void);
|
|
extern void lock_ipi_call_lock(void);
|
|
extern void unlock_ipi_call_lock(void);
|
|
|
|
extern int smp_call_function_mask(cpumask_t mask, void (*func)(void *),
|
|
void *info, int wait);
|
|
|
|
extern u16 __initdata x86_cpu_to_apicid_init[];
|
|
extern u16 __initdata x86_bios_cpu_apicid_init[];
|
|
extern void *x86_cpu_to_apicid_early_ptr;
|
|
extern void *x86_bios_cpu_apicid_early_ptr;
|
|
|
|
DECLARE_PER_CPU(cpumask_t, cpu_sibling_map);
|
|
DECLARE_PER_CPU(cpumask_t, cpu_core_map);
|
|
DECLARE_PER_CPU(u16, cpu_llc_id);
|
|
DECLARE_PER_CPU(u16, x86_cpu_to_apicid);
|
|
DECLARE_PER_CPU(u16, x86_bios_cpu_apicid);
|
|
|
|
static inline int cpu_present_to_apicid(int mps_cpu)
|
|
{
|
|
if (cpu_present(mps_cpu))
|
|
return (int)per_cpu(x86_bios_cpu_apicid, mps_cpu);
|
|
else
|
|
return BAD_APICID;
|
|
}
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
#define SMP_TRAMPOLINE_BASE 0x6000
|
|
|
|
extern int __cpu_disable(void);
|
|
extern void __cpu_die(unsigned int cpu);
|
|
extern void prefill_possible_map(void);
|
|
extern unsigned __cpuinitdata disabled_cpus;
|
|
|
|
#define raw_smp_processor_id() read_pda(cpunumber)
|
|
#define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu)
|
|
|
|
#define stack_smp_processor_id() \
|
|
({ \
|
|
struct thread_info *ti; \
|
|
__asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \
|
|
ti->cpu; \
|
|
})
|
|
|
|
/*
|
|
* On x86 all CPUs are mapped 1:1 to the APIC space. This simplifies
|
|
* scheduling and IPI sending and compresses data structures.
|
|
*/
|
|
static inline int num_booting_cpus(void)
|
|
{
|
|
return cpus_weight(cpu_callout_map);
|
|
}
|
|
|
|
extern void smp_send_reschedule(int cpu);
|
|
|
|
#else /* CONFIG_SMP */
|
|
|
|
extern unsigned int boot_cpu_id;
|
|
#define cpu_physical_id(cpu) boot_cpu_id
|
|
#define stack_smp_processor_id() 0
|
|
|
|
#endif /* !CONFIG_SMP */
|
|
|
|
#define safe_smp_processor_id() smp_processor_id()
|
|
|
|
static __inline int logical_smp_processor_id(void)
|
|
{
|
|
/* we don't want to mark this access volatile - bad code generation */
|
|
return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR));
|
|
}
|
|
|
|
static inline int hard_smp_processor_id(void)
|
|
{
|
|
/* we don't want to mark this access volatile - bad code generation */
|
|
return GET_APIC_ID(*(u32 *)(APIC_BASE + APIC_ID));
|
|
}
|
|
|
|
#endif
|
|
|