2011-11-30 02:46:42 +00:00
|
|
|
/*
|
|
|
|
* processor_idle - idle state cpuidle driver.
|
|
|
|
* Adapted from drivers/idle/intel_idle.c and
|
|
|
|
* drivers/acpi/processor_idle.c
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/moduleparam.h>
|
|
|
|
#include <linux/cpuidle.h>
|
|
|
|
#include <linux/cpu.h>
|
|
|
|
|
|
|
|
#include <asm/paca.h>
|
|
|
|
#include <asm/reg.h>
|
|
|
|
#include <asm/machdep.h>
|
|
|
|
#include <asm/firmware.h>
|
2012-03-28 17:30:02 +00:00
|
|
|
#include <asm/runlatch.h>
|
2011-11-30 02:46:42 +00:00
|
|
|
|
|
|
|
#include "plpar_wrappers.h"
|
|
|
|
#include "pseries.h"
|
|
|
|
|
|
|
|
struct cpuidle_driver pseries_idle_driver = {
|
|
|
|
.name = "pseries_idle",
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
#define MAX_IDLE_STATE_COUNT 2
|
|
|
|
|
|
|
|
static int max_idle_state = MAX_IDLE_STATE_COUNT - 1;
|
|
|
|
static struct cpuidle_device __percpu *pseries_cpuidle_devices;
|
|
|
|
static struct cpuidle_state *cpuidle_state_table;
|
|
|
|
|
|
|
|
void update_smt_snooze_delay(int snooze)
|
|
|
|
{
|
|
|
|
struct cpuidle_driver *drv = cpuidle_get_driver();
|
|
|
|
if (drv)
|
|
|
|
drv->states[0].target_residency = snooze;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void idle_loop_prolog(unsigned long *in_purr, ktime_t *kt_before)
|
|
|
|
{
|
|
|
|
|
|
|
|
*kt_before = ktime_get_real();
|
|
|
|
*in_purr = mfspr(SPRN_PURR);
|
|
|
|
/*
|
|
|
|
* Indicate to the HV that we are idle. Now would be
|
|
|
|
* a good time to find other work to dispatch.
|
|
|
|
*/
|
|
|
|
get_lppaca()->idle = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline s64 idle_loop_epilog(unsigned long in_purr, ktime_t kt_before)
|
|
|
|
{
|
|
|
|
get_lppaca()->wait_state_cycles += mfspr(SPRN_PURR) - in_purr;
|
|
|
|
get_lppaca()->idle = 0;
|
|
|
|
|
|
|
|
return ktime_to_us(ktime_sub(ktime_get_real(), kt_before));
|
|
|
|
}
|
|
|
|
|
|
|
|
static int snooze_loop(struct cpuidle_device *dev,
|
|
|
|
struct cpuidle_driver *drv,
|
|
|
|
int index)
|
|
|
|
{
|
|
|
|
unsigned long in_purr;
|
|
|
|
ktime_t kt_before;
|
|
|
|
unsigned long start_snooze;
|
|
|
|
long snooze = drv->states[0].target_residency;
|
|
|
|
|
|
|
|
idle_loop_prolog(&in_purr, &kt_before);
|
|
|
|
|
|
|
|
if (snooze) {
|
|
|
|
start_snooze = get_tb() + snooze * tb_ticks_per_usec;
|
|
|
|
local_irq_enable();
|
|
|
|
set_thread_flag(TIF_POLLING_NRFLAG);
|
|
|
|
|
|
|
|
while ((snooze < 0) || (get_tb() < start_snooze)) {
|
|
|
|
if (need_resched() || cpu_is_offline(dev->cpu))
|
|
|
|
goto out;
|
|
|
|
ppc64_runlatch_off();
|
|
|
|
HMT_low();
|
|
|
|
HMT_very_low();
|
|
|
|
}
|
|
|
|
|
|
|
|
HMT_medium();
|
|
|
|
clear_thread_flag(TIF_POLLING_NRFLAG);
|
|
|
|
smp_mb();
|
|
|
|
local_irq_disable();
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
HMT_medium();
|
|
|
|
dev->last_residency =
|
|
|
|
(int)idle_loop_epilog(in_purr, kt_before);
|
|
|
|
return index;
|
|
|
|
}
|
|
|
|
|
powerpc: Rework lazy-interrupt handling
The current implementation of lazy interrupts handling has some
issues that this tries to address.
We don't do the various workarounds we need to do when re-enabling
interrupts in some cases such as when returning from an interrupt
and thus we may still lose or get delayed decrementer or doorbell
interrupts.
The current scheme also makes it much harder to handle the external
"edge" interrupts provided by some BookE processors when using the
EPR facility (External Proxy) and the Freescale Hypervisor.
Additionally, we tend to keep interrupts hard disabled in a number
of cases, such as decrementer interrupts, external interrupts, or
when a masked decrementer interrupt is pending. This is sub-optimal.
This is an attempt at fixing it all in one go by reworking the way
we do the lazy interrupt disabling from the ground up.
The base idea is to replace the "hard_enabled" field with a
"irq_happened" field in which we store a bit mask of what interrupt
occurred while soft-disabled.
When re-enabling, either via arch_local_irq_restore() or when returning
from an interrupt, we can now decide what to do by testing bits in that
field.
We then implement replaying of the missed interrupts either by
re-using the existing exception frame (in exception exit case) or via
the creation of a new one from an assembly trampoline (in the
arch_local_irq_enable case).
This removes the need to play with the decrementer to try to create
fake interrupts, among others.
In addition, this adds a few refinements:
- We no longer hard disable decrementer interrupts that occur
while soft-disabled. We now simply bump the decrementer back to max
(on BookS) or leave it stopped (on BookE) and continue with hard interrupts
enabled, which means that we'll potentially get better sample quality from
performance monitor interrupts.
- Timer, decrementer and doorbell interrupts now hard-enable
shortly after removing the source of the interrupt, which means
they no longer run entirely hard disabled. Again, this will improve
perf sample quality.
- On Book3E 64-bit, we now make the performance monitor interrupt
act as an NMI like Book3S (the necessary C code for that to work
appear to already be present in the FSL perf code, notably calling
nmi_enter instead of irq_enter). (This also fixes a bug where BookE
perfmon interrupts could clobber r14 ... oops)
- We could make "masked" decrementer interrupts act as NMIs when doing
timer-based perf sampling to improve the sample quality.
Signed-off-by-yet: Benjamin Herrenschmidt <benh@kernel.crashing.org>
---
v2:
- Add hard-enable to decrementer, timer and doorbells
- Fix CR clobber in masked irq handling on BookE
- Make embedded perf interrupt act as an NMI
- Add a PACA_HAPPENED_EE_EDGE for use by FSL if they want
to retrigger an interrupt without preventing hard-enable
v3:
- Fix or vs. ori bug on Book3E
- Fix enabling of interrupts for some exceptions on Book3E
v4:
- Fix resend of doorbells on return from interrupt on Book3E
v5:
- Rebased on top of my latest series, which involves some significant
rework of some aspects of the patch.
v6:
- 32-bit compile fix
- more compile fixes with various .config combos
- factor out the asm code to soft-disable interrupts
- remove the C wrapper around preempt_schedule_irq
v7:
- Fix a bug with hard irq state tracking on native power7
2012-03-06 07:27:59 +00:00
|
|
|
static void check_and_cede_processor(void)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Interrupts are soft-disabled at this point,
|
|
|
|
* but not hard disabled. So an interrupt might have
|
|
|
|
* occurred before entering NAP, and would be potentially
|
|
|
|
* lost (edge events, decrementer events, etc...) unless
|
|
|
|
* we first hard disable then check.
|
|
|
|
*/
|
|
|
|
hard_irq_disable();
|
2012-06-27 13:13:52 +00:00
|
|
|
if (!lazy_irq_pending())
|
powerpc: Rework lazy-interrupt handling
The current implementation of lazy interrupts handling has some
issues that this tries to address.
We don't do the various workarounds we need to do when re-enabling
interrupts in some cases such as when returning from an interrupt
and thus we may still lose or get delayed decrementer or doorbell
interrupts.
The current scheme also makes it much harder to handle the external
"edge" interrupts provided by some BookE processors when using the
EPR facility (External Proxy) and the Freescale Hypervisor.
Additionally, we tend to keep interrupts hard disabled in a number
of cases, such as decrementer interrupts, external interrupts, or
when a masked decrementer interrupt is pending. This is sub-optimal.
This is an attempt at fixing it all in one go by reworking the way
we do the lazy interrupt disabling from the ground up.
The base idea is to replace the "hard_enabled" field with a
"irq_happened" field in which we store a bit mask of what interrupt
occurred while soft-disabled.
When re-enabling, either via arch_local_irq_restore() or when returning
from an interrupt, we can now decide what to do by testing bits in that
field.
We then implement replaying of the missed interrupts either by
re-using the existing exception frame (in exception exit case) or via
the creation of a new one from an assembly trampoline (in the
arch_local_irq_enable case).
This removes the need to play with the decrementer to try to create
fake interrupts, among others.
In addition, this adds a few refinements:
- We no longer hard disable decrementer interrupts that occur
while soft-disabled. We now simply bump the decrementer back to max
(on BookS) or leave it stopped (on BookE) and continue with hard interrupts
enabled, which means that we'll potentially get better sample quality from
performance monitor interrupts.
- Timer, decrementer and doorbell interrupts now hard-enable
shortly after removing the source of the interrupt, which means
they no longer run entirely hard disabled. Again, this will improve
perf sample quality.
- On Book3E 64-bit, we now make the performance monitor interrupt
act as an NMI like Book3S (the necessary C code for that to work
appear to already be present in the FSL perf code, notably calling
nmi_enter instead of irq_enter). (This also fixes a bug where BookE
perfmon interrupts could clobber r14 ... oops)
- We could make "masked" decrementer interrupts act as NMIs when doing
timer-based perf sampling to improve the sample quality.
Signed-off-by-yet: Benjamin Herrenschmidt <benh@kernel.crashing.org>
---
v2:
- Add hard-enable to decrementer, timer and doorbells
- Fix CR clobber in masked irq handling on BookE
- Make embedded perf interrupt act as an NMI
- Add a PACA_HAPPENED_EE_EDGE for use by FSL if they want
to retrigger an interrupt without preventing hard-enable
v3:
- Fix or vs. ori bug on Book3E
- Fix enabling of interrupts for some exceptions on Book3E
v4:
- Fix resend of doorbells on return from interrupt on Book3E
v5:
- Rebased on top of my latest series, which involves some significant
rework of some aspects of the patch.
v6:
- 32-bit compile fix
- more compile fixes with various .config combos
- factor out the asm code to soft-disable interrupts
- remove the C wrapper around preempt_schedule_irq
v7:
- Fix a bug with hard irq state tracking on native power7
2012-03-06 07:27:59 +00:00
|
|
|
cede_processor();
|
|
|
|
}
|
|
|
|
|
2011-11-30 02:46:42 +00:00
|
|
|
static int dedicated_cede_loop(struct cpuidle_device *dev,
|
|
|
|
struct cpuidle_driver *drv,
|
|
|
|
int index)
|
|
|
|
{
|
|
|
|
unsigned long in_purr;
|
|
|
|
ktime_t kt_before;
|
|
|
|
|
|
|
|
idle_loop_prolog(&in_purr, &kt_before);
|
|
|
|
get_lppaca()->donate_dedicated_cpu = 1;
|
|
|
|
|
|
|
|
ppc64_runlatch_off();
|
|
|
|
HMT_medium();
|
powerpc: Rework lazy-interrupt handling
The current implementation of lazy interrupts handling has some
issues that this tries to address.
We don't do the various workarounds we need to do when re-enabling
interrupts in some cases such as when returning from an interrupt
and thus we may still lose or get delayed decrementer or doorbell
interrupts.
The current scheme also makes it much harder to handle the external
"edge" interrupts provided by some BookE processors when using the
EPR facility (External Proxy) and the Freescale Hypervisor.
Additionally, we tend to keep interrupts hard disabled in a number
of cases, such as decrementer interrupts, external interrupts, or
when a masked decrementer interrupt is pending. This is sub-optimal.
This is an attempt at fixing it all in one go by reworking the way
we do the lazy interrupt disabling from the ground up.
The base idea is to replace the "hard_enabled" field with a
"irq_happened" field in which we store a bit mask of what interrupt
occurred while soft-disabled.
When re-enabling, either via arch_local_irq_restore() or when returning
from an interrupt, we can now decide what to do by testing bits in that
field.
We then implement replaying of the missed interrupts either by
re-using the existing exception frame (in exception exit case) or via
the creation of a new one from an assembly trampoline (in the
arch_local_irq_enable case).
This removes the need to play with the decrementer to try to create
fake interrupts, among others.
In addition, this adds a few refinements:
- We no longer hard disable decrementer interrupts that occur
while soft-disabled. We now simply bump the decrementer back to max
(on BookS) or leave it stopped (on BookE) and continue with hard interrupts
enabled, which means that we'll potentially get better sample quality from
performance monitor interrupts.
- Timer, decrementer and doorbell interrupts now hard-enable
shortly after removing the source of the interrupt, which means
they no longer run entirely hard disabled. Again, this will improve
perf sample quality.
- On Book3E 64-bit, we now make the performance monitor interrupt
act as an NMI like Book3S (the necessary C code for that to work
appear to already be present in the FSL perf code, notably calling
nmi_enter instead of irq_enter). (This also fixes a bug where BookE
perfmon interrupts could clobber r14 ... oops)
- We could make "masked" decrementer interrupts act as NMIs when doing
timer-based perf sampling to improve the sample quality.
Signed-off-by-yet: Benjamin Herrenschmidt <benh@kernel.crashing.org>
---
v2:
- Add hard-enable to decrementer, timer and doorbells
- Fix CR clobber in masked irq handling on BookE
- Make embedded perf interrupt act as an NMI
- Add a PACA_HAPPENED_EE_EDGE for use by FSL if they want
to retrigger an interrupt without preventing hard-enable
v3:
- Fix or vs. ori bug on Book3E
- Fix enabling of interrupts for some exceptions on Book3E
v4:
- Fix resend of doorbells on return from interrupt on Book3E
v5:
- Rebased on top of my latest series, which involves some significant
rework of some aspects of the patch.
v6:
- 32-bit compile fix
- more compile fixes with various .config combos
- factor out the asm code to soft-disable interrupts
- remove the C wrapper around preempt_schedule_irq
v7:
- Fix a bug with hard irq state tracking on native power7
2012-03-06 07:27:59 +00:00
|
|
|
check_and_cede_processor();
|
2011-11-30 02:46:42 +00:00
|
|
|
|
|
|
|
get_lppaca()->donate_dedicated_cpu = 0;
|
|
|
|
dev->last_residency =
|
|
|
|
(int)idle_loop_epilog(in_purr, kt_before);
|
|
|
|
return index;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int shared_cede_loop(struct cpuidle_device *dev,
|
|
|
|
struct cpuidle_driver *drv,
|
|
|
|
int index)
|
|
|
|
{
|
|
|
|
unsigned long in_purr;
|
|
|
|
ktime_t kt_before;
|
|
|
|
|
|
|
|
idle_loop_prolog(&in_purr, &kt_before);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Yield the processor to the hypervisor. We return if
|
|
|
|
* an external interrupt occurs (which are driven prior
|
|
|
|
* to returning here) or if a prod occurs from another
|
|
|
|
* processor. When returning here, external interrupts
|
|
|
|
* are enabled.
|
|
|
|
*/
|
powerpc: Rework lazy-interrupt handling
The current implementation of lazy interrupts handling has some
issues that this tries to address.
We don't do the various workarounds we need to do when re-enabling
interrupts in some cases such as when returning from an interrupt
and thus we may still lose or get delayed decrementer or doorbell
interrupts.
The current scheme also makes it much harder to handle the external
"edge" interrupts provided by some BookE processors when using the
EPR facility (External Proxy) and the Freescale Hypervisor.
Additionally, we tend to keep interrupts hard disabled in a number
of cases, such as decrementer interrupts, external interrupts, or
when a masked decrementer interrupt is pending. This is sub-optimal.
This is an attempt at fixing it all in one go by reworking the way
we do the lazy interrupt disabling from the ground up.
The base idea is to replace the "hard_enabled" field with a
"irq_happened" field in which we store a bit mask of what interrupt
occurred while soft-disabled.
When re-enabling, either via arch_local_irq_restore() or when returning
from an interrupt, we can now decide what to do by testing bits in that
field.
We then implement replaying of the missed interrupts either by
re-using the existing exception frame (in exception exit case) or via
the creation of a new one from an assembly trampoline (in the
arch_local_irq_enable case).
This removes the need to play with the decrementer to try to create
fake interrupts, among others.
In addition, this adds a few refinements:
- We no longer hard disable decrementer interrupts that occur
while soft-disabled. We now simply bump the decrementer back to max
(on BookS) or leave it stopped (on BookE) and continue with hard interrupts
enabled, which means that we'll potentially get better sample quality from
performance monitor interrupts.
- Timer, decrementer and doorbell interrupts now hard-enable
shortly after removing the source of the interrupt, which means
they no longer run entirely hard disabled. Again, this will improve
perf sample quality.
- On Book3E 64-bit, we now make the performance monitor interrupt
act as an NMI like Book3S (the necessary C code for that to work
appear to already be present in the FSL perf code, notably calling
nmi_enter instead of irq_enter). (This also fixes a bug where BookE
perfmon interrupts could clobber r14 ... oops)
- We could make "masked" decrementer interrupts act as NMIs when doing
timer-based perf sampling to improve the sample quality.
Signed-off-by-yet: Benjamin Herrenschmidt <benh@kernel.crashing.org>
---
v2:
- Add hard-enable to decrementer, timer and doorbells
- Fix CR clobber in masked irq handling on BookE
- Make embedded perf interrupt act as an NMI
- Add a PACA_HAPPENED_EE_EDGE for use by FSL if they want
to retrigger an interrupt without preventing hard-enable
v3:
- Fix or vs. ori bug on Book3E
- Fix enabling of interrupts for some exceptions on Book3E
v4:
- Fix resend of doorbells on return from interrupt on Book3E
v5:
- Rebased on top of my latest series, which involves some significant
rework of some aspects of the patch.
v6:
- 32-bit compile fix
- more compile fixes with various .config combos
- factor out the asm code to soft-disable interrupts
- remove the C wrapper around preempt_schedule_irq
v7:
- Fix a bug with hard irq state tracking on native power7
2012-03-06 07:27:59 +00:00
|
|
|
check_and_cede_processor();
|
2011-11-30 02:46:42 +00:00
|
|
|
|
|
|
|
dev->last_residency =
|
|
|
|
(int)idle_loop_epilog(in_purr, kt_before);
|
|
|
|
return index;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* States for dedicated partition case.
|
|
|
|
*/
|
|
|
|
static struct cpuidle_state dedicated_states[MAX_IDLE_STATE_COUNT] = {
|
|
|
|
{ /* Snooze */
|
|
|
|
.name = "snooze",
|
|
|
|
.desc = "snooze",
|
|
|
|
.flags = CPUIDLE_FLAG_TIME_VALID,
|
|
|
|
.exit_latency = 0,
|
|
|
|
.target_residency = 0,
|
|
|
|
.enter = &snooze_loop },
|
|
|
|
{ /* CEDE */
|
|
|
|
.name = "CEDE",
|
|
|
|
.desc = "CEDE",
|
|
|
|
.flags = CPUIDLE_FLAG_TIME_VALID,
|
|
|
|
.exit_latency = 1,
|
|
|
|
.target_residency = 10,
|
|
|
|
.enter = &dedicated_cede_loop },
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* States for shared partition case.
|
|
|
|
*/
|
|
|
|
static struct cpuidle_state shared_states[MAX_IDLE_STATE_COUNT] = {
|
|
|
|
{ /* Shared Cede */
|
|
|
|
.name = "Shared Cede",
|
|
|
|
.desc = "Shared Cede",
|
|
|
|
.flags = CPUIDLE_FLAG_TIME_VALID,
|
|
|
|
.exit_latency = 0,
|
|
|
|
.target_residency = 0,
|
|
|
|
.enter = &shared_cede_loop },
|
|
|
|
};
|
|
|
|
|
|
|
|
int pseries_notify_cpuidle_add_cpu(int cpu)
|
|
|
|
{
|
|
|
|
struct cpuidle_device *dev =
|
|
|
|
per_cpu_ptr(pseries_cpuidle_devices, cpu);
|
|
|
|
if (dev && cpuidle_get_driver()) {
|
|
|
|
cpuidle_disable_device(dev);
|
|
|
|
cpuidle_enable_device(dev);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* pseries_cpuidle_driver_init()
|
|
|
|
*/
|
|
|
|
static int pseries_cpuidle_driver_init(void)
|
|
|
|
{
|
|
|
|
int idle_state;
|
|
|
|
struct cpuidle_driver *drv = &pseries_idle_driver;
|
|
|
|
|
|
|
|
drv->state_count = 0;
|
|
|
|
|
|
|
|
for (idle_state = 0; idle_state < MAX_IDLE_STATE_COUNT; ++idle_state) {
|
|
|
|
|
|
|
|
if (idle_state > max_idle_state)
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* is the state not enabled? */
|
|
|
|
if (cpuidle_state_table[idle_state].enter == NULL)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
drv->states[drv->state_count] = /* structure copy */
|
|
|
|
cpuidle_state_table[idle_state];
|
|
|
|
|
|
|
|
if (cpuidle_state_table == dedicated_states)
|
|
|
|
drv->states[drv->state_count].target_residency =
|
|
|
|
__get_cpu_var(smt_snooze_delay);
|
|
|
|
|
|
|
|
drv->state_count += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* pseries_idle_devices_uninit(void)
|
|
|
|
* unregister cpuidle devices and de-allocate memory
|
|
|
|
*/
|
|
|
|
static void pseries_idle_devices_uninit(void)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
struct cpuidle_device *dev;
|
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
dev = per_cpu_ptr(pseries_cpuidle_devices, i);
|
|
|
|
cpuidle_unregister_device(dev);
|
|
|
|
}
|
|
|
|
|
|
|
|
free_percpu(pseries_cpuidle_devices);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* pseries_idle_devices_init()
|
|
|
|
* allocate, initialize and register cpuidle device
|
|
|
|
*/
|
|
|
|
static int pseries_idle_devices_init(void)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
struct cpuidle_driver *drv = &pseries_idle_driver;
|
|
|
|
struct cpuidle_device *dev;
|
|
|
|
|
|
|
|
pseries_cpuidle_devices = alloc_percpu(struct cpuidle_device);
|
|
|
|
if (pseries_cpuidle_devices == NULL)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
dev = per_cpu_ptr(pseries_cpuidle_devices, i);
|
|
|
|
dev->state_count = drv->state_count;
|
|
|
|
dev->cpu = i;
|
|
|
|
if (cpuidle_register_device(dev)) {
|
|
|
|
printk(KERN_DEBUG \
|
|
|
|
"cpuidle_register_device %d failed!\n", i);
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* pseries_idle_probe()
|
|
|
|
* Choose state table for shared versus dedicated partition
|
|
|
|
*/
|
|
|
|
static int pseries_idle_probe(void)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (!firmware_has_feature(FW_FEATURE_SPLPAR))
|
|
|
|
return -ENODEV;
|
|
|
|
|
2011-11-30 02:47:03 +00:00
|
|
|
if (cpuidle_disable != IDLE_NO_OVERRIDE)
|
|
|
|
return -ENODEV;
|
|
|
|
|
2011-11-30 02:46:42 +00:00
|
|
|
if (max_idle_state == 0) {
|
|
|
|
printk(KERN_DEBUG "pseries processor idle disabled.\n");
|
|
|
|
return -EPERM;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (get_lppaca()->shared_proc)
|
|
|
|
cpuidle_state_table = shared_states;
|
|
|
|
else
|
|
|
|
cpuidle_state_table = dedicated_states;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __init pseries_processor_idle_init(void)
|
|
|
|
{
|
|
|
|
int retval;
|
|
|
|
|
|
|
|
retval = pseries_idle_probe();
|
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
pseries_cpuidle_driver_init();
|
|
|
|
retval = cpuidle_register_driver(&pseries_idle_driver);
|
|
|
|
if (retval) {
|
|
|
|
printk(KERN_DEBUG "Registration of pseries driver failed.\n");
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
retval = pseries_idle_devices_init();
|
|
|
|
if (retval) {
|
|
|
|
pseries_idle_devices_uninit();
|
|
|
|
cpuidle_unregister_driver(&pseries_idle_driver);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
printk(KERN_DEBUG "pseries_idle_driver registered\n");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __exit pseries_processor_idle_exit(void)
|
|
|
|
{
|
|
|
|
|
|
|
|
pseries_idle_devices_uninit();
|
|
|
|
cpuidle_unregister_driver(&pseries_idle_driver);
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
module_init(pseries_processor_idle_init);
|
|
|
|
module_exit(pseries_processor_idle_exit);
|
|
|
|
|
|
|
|
MODULE_AUTHOR("Deepthi Dharwar <deepthi@linux.vnet.ibm.com>");
|
|
|
|
MODULE_DESCRIPTION("Cpuidle driver for POWER");
|
|
|
|
MODULE_LICENSE("GPL");
|