forked from Minki/linux
783502719c
We call cpu_cluster_pm_enter for dev->cpu == 0 only, but
cpu_cluster_pm_exit called without that check.
Because of that unhandled page fault may happen:
[ 3.803405] Unable to handle kernel paging request at virtual address 00002500
[ 3.810974] pgd = c0004000
[ 3.813812] [00002500] *pgd=00000000
[ 3.817596] Internal error: Oops: 5 [#1] SMP ARM
[ 3.822418] Modules linked in:
[ 3.825653] CPU: 1 PID: 0 Comm: swapper/1 Not tainted 3.11.0-rc6+ #21
[ 3.832397] task: ed86ef40 ti: ed896000 task.ti: ed896000
[ 3.838073] PC is at irq_notifier+0x234/0x25c
[ 3.842651] LR is at irq_notifier+0x218/0x25c
[ 3.847229] pc : [<c0029ed8>] lr : [<c0029ebc>] psr: 80000193
[ 3.847229] sp : ed897ee8 ip : 00000005 fp : 00000001
[ 3.859283] r10: c0b395f0 r9 : c0b30594 r8 : c0b8c2ac
[ 3.864776] r7 : ffffffff r6 : 00000000 r5 : 00000005 r4 : 00000000
[ 3.871643] r3 : 00002500 r2 : 00000000 r1 : 00000005 r0 : 44302244
[ 3.878479] Flags: Nzcv IRQs off FIQs on Mode SVC_32 ISA ARM Segment kernel
[ 3.886260] Control: 10c5387d Table: 8000404a DAC: 00000015
[ 3.892272] Process swapper/1 (pid: 0, stack limit = 0xed896240)
[ 3.898590] Stack: (0xed897ee8 to 0xed898000)
[ 3.903167] 7ee0: c0979c3a
00000001 ed897ef8 ed896000 c0014f7c 00000000
[ 3.911743] 7f00: 00000005 00000000 ffffffff c0b8c2ac c0b395f0 c077c04c c0c94b48 c0b3953c
[ 3.920318] 7f20: c0bcd928 00000002 c0b39524 c00cfad8 00000000 ffffffff 00000000 c00cfb10
[ 3.928924] 7f40: c14e62c0 c002c1c8 c002c0ac c14e62c0 00000002 e251c37d 00000000 c0b39548
[ 3.937499] 7f60: c0b395f0 c05a1bc4 e251c37d 00000000 00000005 c05a3870 edc90380 edc90380
[ 3.946105] 7f80: edc90394 c14e62c0 c0b39548 00000002 c0784064 c05a3c78 c0b395e0 c14e62c0
[ 3.954681] 7fa0: 00000002 c0b39548 c0bc9db8 00000000 00000001 c05a1dc0 ed896000 00000015
[ 3.963287] 7fc0: c0bc9db8 ed896000 8000406a c0b30594 c0784064 c000e504 00000746 c007a528
[ 3.971862] 7fe0: 00000001 0000001d 600001d3 c0bcc004 00000000 800086c4 ee0aa6a7 d2aabaa9
[ 3.980499] [<c0029ed8>] (irq_notifier+0x234/0x25c) from [<c077c04c>] (notifier_call_chain+0x38/0x68)
[ 3.990173] [<c077c04c>] (notifier_call_chain+0x38/0x68) from [<c00cfad8>] (cpu_pm_notify+0x20/0x38)
[ 3.999786] [<c00cfad8>] (cpu_pm_notify+0x20/0x38) from [<c00cfb10>] (cpu_cluster_pm_exit+0x20/0x50)
[ 4.009399] [<c00cfb10>] (cpu_cluster_pm_exit+0x20/0x50) from [<c002c1c8>] (omap_enter_idle_coupled+0x11c/0x14c)
[ 4.020111] [<c002c1c8>] (omap_enter_idle_coupled+0x11c/0x14c) from [<c05a1bc4>] (cpuidle_enter_state+0x40/0xec)
[ 4.030822] [<c05a1bc4>] (cpuidle_enter_state+0x40/0xec) from [<c05a3c78>] (cpuidle_enter_state_coupled+0x1f4/0x240)
[ 4.041870] [<c05a3c78>] (cpuidle_enter_state_coupled+0x1f4/0x240) from [<c05a1dc0>] (cpuidle_idle_call+0x150/0x228)
[ 4.052947] [<c05a1dc0>] (cpuidle_idle_call+0x150/0x228) from [<c000e504>] (arch_cpu_idle+0x8/0x38)
[ 4.062499] [<c000e504>] (arch_cpu_idle+0x8/0x38) from [<c007a528>] (cpu_startup_entry+0x178/0x1e4)
[ 4.071990] [<c007a528>] (cpu_startup_entry+0x178/0x1e4) from [<800086c4>] (0x800086c4)
[ 4.080383] Code: e5922288 03a03b0a 13a03c25 e0823003 (e5932000)
[ 4.086791] ---[ end trace d83954a84a6fa69e ]---
It is supposed that sar_base is initialized in irq_save_context, which
is called on CPU_CLUSTER_PM_ENTER notification. If this notification
has been missed and CPU_CLUSTER_PM_EXIT is received sar_base is NULL.
Fix it by calling CPU_CLUSTER_PM_{ENTER,EXIT} under the same condition.
Signed-off-by: Vladimir Murzin <murzin.v@gmail.com>
Acked-by: Kevin Hilman <khilman@linaro.org>
Signed-off-by: Tony Lindgren <tony@atomide.com>
218 lines
5.4 KiB
C
218 lines
5.4 KiB
C
/*
|
|
* OMAP4+ CPU idle Routines
|
|
*
|
|
* Copyright (C) 2011-2013 Texas Instruments, Inc.
|
|
* Santosh Shilimkar <santosh.shilimkar@ti.com>
|
|
* Rajendra Nayak <rnayak@ti.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/cpuidle.h>
|
|
#include <linux/cpu_pm.h>
|
|
#include <linux/export.h>
|
|
|
|
#include <asm/cpuidle.h>
|
|
#include <asm/proc-fns.h>
|
|
|
|
#include "common.h"
|
|
#include "pm.h"
|
|
#include "prm.h"
|
|
#include "clockdomain.h"
|
|
|
|
/* Machine specific information */
|
|
struct idle_statedata {
|
|
u32 cpu_state;
|
|
u32 mpu_logic_state;
|
|
u32 mpu_state;
|
|
};
|
|
|
|
static struct idle_statedata omap4_idle_data[] = {
|
|
{
|
|
.cpu_state = PWRDM_POWER_ON,
|
|
.mpu_state = PWRDM_POWER_ON,
|
|
.mpu_logic_state = PWRDM_POWER_RET,
|
|
},
|
|
{
|
|
.cpu_state = PWRDM_POWER_OFF,
|
|
.mpu_state = PWRDM_POWER_RET,
|
|
.mpu_logic_state = PWRDM_POWER_RET,
|
|
},
|
|
{
|
|
.cpu_state = PWRDM_POWER_OFF,
|
|
.mpu_state = PWRDM_POWER_RET,
|
|
.mpu_logic_state = PWRDM_POWER_OFF,
|
|
},
|
|
};
|
|
|
|
static struct powerdomain *mpu_pd, *cpu_pd[NR_CPUS];
|
|
static struct clockdomain *cpu_clkdm[NR_CPUS];
|
|
|
|
static atomic_t abort_barrier;
|
|
static bool cpu_done[NR_CPUS];
|
|
static struct idle_statedata *state_ptr = &omap4_idle_data[0];
|
|
|
|
/* Private functions */
|
|
|
|
/**
|
|
* omap_enter_idle_[simple/coupled] - OMAP4PLUS cpuidle entry functions
|
|
* @dev: cpuidle device
|
|
* @drv: cpuidle driver
|
|
* @index: the index of state to be entered
|
|
*
|
|
* Called from the CPUidle framework to program the device to the
|
|
* specified low power state selected by the governor.
|
|
* Returns the amount of time spent in the low power state.
|
|
*/
|
|
static int omap_enter_idle_simple(struct cpuidle_device *dev,
|
|
struct cpuidle_driver *drv,
|
|
int index)
|
|
{
|
|
omap_do_wfi();
|
|
return index;
|
|
}
|
|
|
|
static int omap_enter_idle_coupled(struct cpuidle_device *dev,
|
|
struct cpuidle_driver *drv,
|
|
int index)
|
|
{
|
|
struct idle_statedata *cx = state_ptr + index;
|
|
|
|
/*
|
|
* CPU0 has to wait and stay ON until CPU1 is OFF state.
|
|
* This is necessary to honour hardware recommondation
|
|
* of triggeing all the possible low power modes once CPU1 is
|
|
* out of coherency and in OFF mode.
|
|
*/
|
|
if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
|
|
while (pwrdm_read_pwrst(cpu_pd[1]) != PWRDM_POWER_OFF) {
|
|
cpu_relax();
|
|
|
|
/*
|
|
* CPU1 could have already entered & exited idle
|
|
* without hitting off because of a wakeup
|
|
* or a failed attempt to hit off mode. Check for
|
|
* that here, otherwise we could spin forever
|
|
* waiting for CPU1 off.
|
|
*/
|
|
if (cpu_done[1])
|
|
goto fail;
|
|
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Call idle CPU PM enter notifier chain so that
|
|
* VFP and per CPU interrupt context is saved.
|
|
*/
|
|
cpu_pm_enter();
|
|
|
|
if (dev->cpu == 0) {
|
|
pwrdm_set_logic_retst(mpu_pd, cx->mpu_logic_state);
|
|
omap_set_pwrdm_state(mpu_pd, cx->mpu_state);
|
|
|
|
/*
|
|
* Call idle CPU cluster PM enter notifier chain
|
|
* to save GIC and wakeupgen context.
|
|
*/
|
|
if ((cx->mpu_state == PWRDM_POWER_RET) &&
|
|
(cx->mpu_logic_state == PWRDM_POWER_OFF))
|
|
cpu_cluster_pm_enter();
|
|
}
|
|
|
|
omap4_enter_lowpower(dev->cpu, cx->cpu_state);
|
|
cpu_done[dev->cpu] = true;
|
|
|
|
/* Wakeup CPU1 only if it is not offlined */
|
|
if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
|
|
clkdm_wakeup(cpu_clkdm[1]);
|
|
omap_set_pwrdm_state(cpu_pd[1], PWRDM_POWER_ON);
|
|
clkdm_allow_idle(cpu_clkdm[1]);
|
|
}
|
|
|
|
/*
|
|
* Call idle CPU PM exit notifier chain to restore
|
|
* VFP and per CPU IRQ context.
|
|
*/
|
|
cpu_pm_exit();
|
|
|
|
/*
|
|
* Call idle CPU cluster PM exit notifier chain
|
|
* to restore GIC and wakeupgen context.
|
|
*/
|
|
if (dev->cpu == 0 && (cx->mpu_state == PWRDM_POWER_RET) &&
|
|
(cx->mpu_logic_state == PWRDM_POWER_OFF))
|
|
cpu_cluster_pm_exit();
|
|
|
|
fail:
|
|
cpuidle_coupled_parallel_barrier(dev, &abort_barrier);
|
|
cpu_done[dev->cpu] = false;
|
|
|
|
return index;
|
|
}
|
|
|
|
static struct cpuidle_driver omap4_idle_driver = {
|
|
.name = "omap4_idle",
|
|
.owner = THIS_MODULE,
|
|
.states = {
|
|
{
|
|
/* C1 - CPU0 ON + CPU1 ON + MPU ON */
|
|
.exit_latency = 2 + 2,
|
|
.target_residency = 5,
|
|
.flags = CPUIDLE_FLAG_TIME_VALID,
|
|
.enter = omap_enter_idle_simple,
|
|
.name = "C1",
|
|
.desc = "CPUx ON, MPUSS ON"
|
|
},
|
|
{
|
|
/* C2 - CPU0 OFF + CPU1 OFF + MPU CSWR */
|
|
.exit_latency = 328 + 440,
|
|
.target_residency = 960,
|
|
.flags = CPUIDLE_FLAG_TIME_VALID | CPUIDLE_FLAG_COUPLED |
|
|
CPUIDLE_FLAG_TIMER_STOP,
|
|
.enter = omap_enter_idle_coupled,
|
|
.name = "C2",
|
|
.desc = "CPUx OFF, MPUSS CSWR",
|
|
},
|
|
{
|
|
/* C3 - CPU0 OFF + CPU1 OFF + MPU OSWR */
|
|
.exit_latency = 460 + 518,
|
|
.target_residency = 1100,
|
|
.flags = CPUIDLE_FLAG_TIME_VALID | CPUIDLE_FLAG_COUPLED |
|
|
CPUIDLE_FLAG_TIMER_STOP,
|
|
.enter = omap_enter_idle_coupled,
|
|
.name = "C3",
|
|
.desc = "CPUx OFF, MPUSS OSWR",
|
|
},
|
|
},
|
|
.state_count = ARRAY_SIZE(omap4_idle_data),
|
|
.safe_state_index = 0,
|
|
};
|
|
|
|
/* Public functions */
|
|
|
|
/**
|
|
* omap4_idle_init - Init routine for OMAP4+ idle
|
|
*
|
|
* Registers the OMAP4+ specific cpuidle driver to the cpuidle
|
|
* framework with the valid set of states.
|
|
*/
|
|
int __init omap4_idle_init(void)
|
|
{
|
|
mpu_pd = pwrdm_lookup("mpu_pwrdm");
|
|
cpu_pd[0] = pwrdm_lookup("cpu0_pwrdm");
|
|
cpu_pd[1] = pwrdm_lookup("cpu1_pwrdm");
|
|
if ((!mpu_pd) || (!cpu_pd[0]) || (!cpu_pd[1]))
|
|
return -ENODEV;
|
|
|
|
cpu_clkdm[0] = clkdm_lookup("mpu0_clkdm");
|
|
cpu_clkdm[1] = clkdm_lookup("mpu1_clkdm");
|
|
if (!cpu_clkdm[0] || !cpu_clkdm[1])
|
|
return -ENODEV;
|
|
|
|
return cpuidle_register(&omap4_idle_driver, cpu_online_mask);
|
|
}
|