bea248fb30
The iSeries code keeps a pointer to the ItLpQueue in its paca struct. But all these pointers end up pointing to the one place, ie. xItLpQueue. So remove the pointer from the paca struct and just refer to xItLpQueue directly where needed. The only complication is that the spread_lpevents logic was implemented by having a NULL lpqueue pointer in the paca on CPUs that weren't supposed to process events. Instead we just compare the spread_lpevents value to the processor id to get the same behaviour. Signed-off-by: Michael Ellerman <michael@ellerman.id.au> Acked-by: Stephen Rothwell <sfr@canb.auug.org.au> Signed-off-by: Paul Mackerras <paulus@samba.org>
384 lines
8.3 KiB
C
384 lines
8.3 KiB
C
/*
|
|
* Idle daemon for PowerPC. Idle daemon will handle any action
|
|
* that needs to be taken when the system becomes idle.
|
|
*
|
|
* Originally Written by Cort Dougan (cort@cs.nmt.edu)
|
|
*
|
|
* iSeries supported added by Mike Corrigan <mikejc@us.ibm.com>
|
|
*
|
|
* Additional shared processor, SMT, and firmware support
|
|
* Copyright (c) 2003 Dave Engebretsen <engebret@us.ibm.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*/
|
|
|
|
#include <linux/config.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/cpu.h>
|
|
#include <linux/module.h>
|
|
#include <linux/sysctl.h>
|
|
#include <linux/smp.h>
|
|
|
|
#include <asm/system.h>
|
|
#include <asm/processor.h>
|
|
#include <asm/mmu.h>
|
|
#include <asm/cputable.h>
|
|
#include <asm/time.h>
|
|
#include <asm/iSeries/HvCall.h>
|
|
#include <asm/iSeries/ItLpQueue.h>
|
|
#include <asm/plpar_wrappers.h>
|
|
#include <asm/systemcfg.h>
|
|
|
|
extern void power4_idle(void);
|
|
|
|
static int (*idle_loop)(void);
|
|
|
|
#ifdef CONFIG_PPC_ISERIES
|
|
static unsigned long maxYieldTime = 0;
|
|
static unsigned long minYieldTime = 0xffffffffffffffffUL;
|
|
|
|
static inline void process_iSeries_events(void)
|
|
{
|
|
asm volatile ("li 0,0x5555; sc" : : : "r0", "r3");
|
|
}
|
|
|
|
static void yield_shared_processor(void)
|
|
{
|
|
unsigned long tb;
|
|
unsigned long yieldTime;
|
|
|
|
HvCall_setEnabledInterrupts(HvCall_MaskIPI |
|
|
HvCall_MaskLpEvent |
|
|
HvCall_MaskLpProd |
|
|
HvCall_MaskTimeout);
|
|
|
|
tb = get_tb();
|
|
/* Compute future tb value when yield should expire */
|
|
HvCall_yieldProcessor(HvCall_YieldTimed, tb+tb_ticks_per_jiffy);
|
|
|
|
yieldTime = get_tb() - tb;
|
|
if (yieldTime > maxYieldTime)
|
|
maxYieldTime = yieldTime;
|
|
|
|
if (yieldTime < minYieldTime)
|
|
minYieldTime = yieldTime;
|
|
|
|
/*
|
|
* The decrementer stops during the yield. Force a fake decrementer
|
|
* here and let the timer_interrupt code sort out the actual time.
|
|
*/
|
|
get_paca()->lppaca.int_dword.fields.decr_int = 1;
|
|
process_iSeries_events();
|
|
}
|
|
|
|
static int iSeries_idle(void)
|
|
{
|
|
struct paca_struct *lpaca;
|
|
long oldval;
|
|
|
|
/* ensure iSeries run light will be out when idle */
|
|
ppc64_runlatch_off();
|
|
|
|
lpaca = get_paca();
|
|
|
|
while (1) {
|
|
if (lpaca->lppaca.shared_proc) {
|
|
if (ItLpQueue_isLpIntPending(&xItLpQueue))
|
|
process_iSeries_events();
|
|
if (!need_resched())
|
|
yield_shared_processor();
|
|
} else {
|
|
oldval = test_and_clear_thread_flag(TIF_NEED_RESCHED);
|
|
|
|
if (!oldval) {
|
|
set_thread_flag(TIF_POLLING_NRFLAG);
|
|
|
|
while (!need_resched()) {
|
|
HMT_medium();
|
|
if (ItLpQueue_isLpIntPending(&xItLpQueue))
|
|
process_iSeries_events();
|
|
HMT_low();
|
|
}
|
|
|
|
HMT_medium();
|
|
clear_thread_flag(TIF_POLLING_NRFLAG);
|
|
} else {
|
|
set_need_resched();
|
|
}
|
|
}
|
|
|
|
ppc64_runlatch_on();
|
|
schedule();
|
|
ppc64_runlatch_off();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#else
|
|
|
|
static int default_idle(void)
|
|
{
|
|
long oldval;
|
|
unsigned int cpu = smp_processor_id();
|
|
|
|
while (1) {
|
|
oldval = test_and_clear_thread_flag(TIF_NEED_RESCHED);
|
|
|
|
if (!oldval) {
|
|
set_thread_flag(TIF_POLLING_NRFLAG);
|
|
|
|
while (!need_resched() && !cpu_is_offline(cpu)) {
|
|
barrier();
|
|
/*
|
|
* Go into low thread priority and possibly
|
|
* low power mode.
|
|
*/
|
|
HMT_low();
|
|
HMT_very_low();
|
|
}
|
|
|
|
HMT_medium();
|
|
clear_thread_flag(TIF_POLLING_NRFLAG);
|
|
} else {
|
|
set_need_resched();
|
|
}
|
|
|
|
schedule();
|
|
if (cpu_is_offline(cpu) && system_state == SYSTEM_RUNNING)
|
|
cpu_die();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_PPC_PSERIES
|
|
|
|
DECLARE_PER_CPU(unsigned long, smt_snooze_delay);
|
|
|
|
int dedicated_idle(void)
|
|
{
|
|
long oldval;
|
|
struct paca_struct *lpaca = get_paca(), *ppaca;
|
|
unsigned long start_snooze;
|
|
unsigned long *smt_snooze_delay = &__get_cpu_var(smt_snooze_delay);
|
|
unsigned int cpu = smp_processor_id();
|
|
|
|
ppaca = &paca[cpu ^ 1];
|
|
|
|
while (1) {
|
|
/*
|
|
* Indicate to the HV that we are idle. Now would be
|
|
* a good time to find other work to dispatch.
|
|
*/
|
|
lpaca->lppaca.idle = 1;
|
|
|
|
oldval = test_and_clear_thread_flag(TIF_NEED_RESCHED);
|
|
if (!oldval) {
|
|
set_thread_flag(TIF_POLLING_NRFLAG);
|
|
start_snooze = __get_tb() +
|
|
*smt_snooze_delay * tb_ticks_per_usec;
|
|
while (!need_resched() && !cpu_is_offline(cpu)) {
|
|
/*
|
|
* Go into low thread priority and possibly
|
|
* low power mode.
|
|
*/
|
|
HMT_low();
|
|
HMT_very_low();
|
|
|
|
if (*smt_snooze_delay == 0 ||
|
|
__get_tb() < start_snooze)
|
|
continue;
|
|
|
|
HMT_medium();
|
|
|
|
if (!(ppaca->lppaca.idle)) {
|
|
local_irq_disable();
|
|
|
|
/*
|
|
* We are about to sleep the thread
|
|
* and so wont be polling any
|
|
* more.
|
|
*/
|
|
clear_thread_flag(TIF_POLLING_NRFLAG);
|
|
|
|
/*
|
|
* SMT dynamic mode. Cede will result
|
|
* in this thread going dormant, if the
|
|
* partner thread is still doing work.
|
|
* Thread wakes up if partner goes idle,
|
|
* an interrupt is presented, or a prod
|
|
* occurs. Returning from the cede
|
|
* enables external interrupts.
|
|
*/
|
|
if (!need_resched())
|
|
cede_processor();
|
|
else
|
|
local_irq_enable();
|
|
} else {
|
|
/*
|
|
* Give the HV an opportunity at the
|
|
* processor, since we are not doing
|
|
* any work.
|
|
*/
|
|
poll_pending();
|
|
}
|
|
}
|
|
|
|
clear_thread_flag(TIF_POLLING_NRFLAG);
|
|
} else {
|
|
set_need_resched();
|
|
}
|
|
|
|
HMT_medium();
|
|
lpaca->lppaca.idle = 0;
|
|
schedule();
|
|
if (cpu_is_offline(cpu) && system_state == SYSTEM_RUNNING)
|
|
cpu_die();
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int shared_idle(void)
|
|
{
|
|
struct paca_struct *lpaca = get_paca();
|
|
unsigned int cpu = smp_processor_id();
|
|
|
|
while (1) {
|
|
/*
|
|
* Indicate to the HV that we are idle. Now would be
|
|
* a good time to find other work to dispatch.
|
|
*/
|
|
lpaca->lppaca.idle = 1;
|
|
|
|
while (!need_resched() && !cpu_is_offline(cpu)) {
|
|
local_irq_disable();
|
|
|
|
/*
|
|
* Yield the processor to the hypervisor. We return if
|
|
* an external interrupt occurs (which are driven prior
|
|
* to returning here) or if a prod occurs from another
|
|
* processor. When returning here, external interrupts
|
|
* are enabled.
|
|
*
|
|
* Check need_resched() again with interrupts disabled
|
|
* to avoid a race.
|
|
*/
|
|
if (!need_resched())
|
|
cede_processor();
|
|
else
|
|
local_irq_enable();
|
|
}
|
|
|
|
HMT_medium();
|
|
lpaca->lppaca.idle = 0;
|
|
schedule();
|
|
if (cpu_is_offline(smp_processor_id()) &&
|
|
system_state == SYSTEM_RUNNING)
|
|
cpu_die();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_PPC_PSERIES */
|
|
|
|
static int native_idle(void)
|
|
{
|
|
while(1) {
|
|
/* check CPU type here */
|
|
if (!need_resched())
|
|
power4_idle();
|
|
if (need_resched())
|
|
schedule();
|
|
|
|
if (cpu_is_offline(raw_smp_processor_id()) &&
|
|
system_state == SYSTEM_RUNNING)
|
|
cpu_die();
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_PPC_ISERIES */
|
|
|
|
void cpu_idle(void)
|
|
{
|
|
idle_loop();
|
|
}
|
|
|
|
int powersave_nap;
|
|
|
|
#ifdef CONFIG_SYSCTL
|
|
/*
|
|
* Register the sysctl to set/clear powersave_nap.
|
|
*/
|
|
static ctl_table powersave_nap_ctl_table[]={
|
|
{
|
|
.ctl_name = KERN_PPC_POWERSAVE_NAP,
|
|
.procname = "powersave-nap",
|
|
.data = &powersave_nap,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_dointvec,
|
|
},
|
|
{ 0, },
|
|
};
|
|
static ctl_table powersave_nap_sysctl_root[] = {
|
|
{ 1, "kernel", NULL, 0, 0755, powersave_nap_ctl_table, },
|
|
{ 0,},
|
|
};
|
|
|
|
static int __init
|
|
register_powersave_nap_sysctl(void)
|
|
{
|
|
register_sysctl_table(powersave_nap_sysctl_root, 0);
|
|
|
|
return 0;
|
|
}
|
|
__initcall(register_powersave_nap_sysctl);
|
|
#endif
|
|
|
|
int idle_setup(void)
|
|
{
|
|
/*
|
|
* Move that junk to each platform specific file, eventually define
|
|
* a pSeries_idle for shared processor stuff
|
|
*/
|
|
#ifdef CONFIG_PPC_ISERIES
|
|
idle_loop = iSeries_idle;
|
|
return 1;
|
|
#else
|
|
idle_loop = default_idle;
|
|
#endif
|
|
#ifdef CONFIG_PPC_PSERIES
|
|
if (systemcfg->platform & PLATFORM_PSERIES) {
|
|
if (cur_cpu_spec->firmware_features & FW_FEATURE_SPLPAR) {
|
|
if (get_paca()->lppaca.shared_proc) {
|
|
printk(KERN_INFO "Using shared processor idle loop\n");
|
|
idle_loop = shared_idle;
|
|
} else {
|
|
printk(KERN_INFO "Using dedicated idle loop\n");
|
|
idle_loop = dedicated_idle;
|
|
}
|
|
} else {
|
|
printk(KERN_INFO "Using default idle loop\n");
|
|
idle_loop = default_idle;
|
|
}
|
|
}
|
|
#endif /* CONFIG_PPC_PSERIES */
|
|
#ifndef CONFIG_PPC_ISERIES
|
|
if (systemcfg->platform == PLATFORM_POWERMAC ||
|
|
systemcfg->platform == PLATFORM_MAPLE) {
|
|
printk(KERN_INFO "Using native/NAP idle loop\n");
|
|
idle_loop = native_idle;
|
|
}
|
|
#endif /* CONFIG_PPC_ISERIES */
|
|
|
|
return 1;
|
|
}
|