2011-02-14 07:33:10 +00:00
|
|
|
/* linux/arch/arm/mach-exynos4/platsmp.c
|
2010-07-26 12:08:52 +00:00
|
|
|
*
|
2011-02-14 07:33:10 +00:00
|
|
|
* Copyright (c) 2010-2011 Samsung Electronics Co., Ltd.
|
|
|
|
* http://www.samsung.com
|
2010-07-26 12:08:52 +00:00
|
|
|
*
|
|
|
|
* Cloned from linux/arch/arm/mach-vexpress/platsmp.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 2002 ARM Ltd.
|
|
|
|
* All Rights Reserved
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/delay.h>
|
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/jiffies.h>
|
|
|
|
#include <linux/smp.h>
|
|
|
|
#include <linux/io.h>
|
|
|
|
|
|
|
|
#include <asm/cacheflush.h>
|
2012-01-20 11:01:12 +00:00
|
|
|
#include <asm/smp_plat.h>
|
2010-07-26 12:08:52 +00:00
|
|
|
#include <asm/smp_scu.h>
|
2012-12-11 04:58:43 +00:00
|
|
|
#include <asm/firmware.h>
|
2010-07-26 12:08:52 +00:00
|
|
|
|
|
|
|
#include <mach/hardware.h>
|
|
|
|
|
2011-08-20 04:41:21 +00:00
|
|
|
#include <plat/cpu.h>
|
|
|
|
|
2011-09-08 12:15:22 +00:00
|
|
|
#include "common.h"
|
2013-12-18 19:06:56 +00:00
|
|
|
#include "regs-pmu.h"
|
2011-09-08 12:15:22 +00:00
|
|
|
|
2011-02-14 07:33:10 +00:00
|
|
|
extern void exynos4_secondary_startup(void);
|
2010-07-26 12:08:52 +00:00
|
|
|
|
2012-11-24 02:13:48 +00:00
|
|
|
static inline void __iomem *cpu_boot_reg_base(void)
|
|
|
|
{
|
|
|
|
if (soc_is_exynos4210() && samsung_rev() == EXYNOS4210_REV_1_1)
|
|
|
|
return S5P_INFORM5;
|
|
|
|
return S5P_VA_SYSRAM;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __iomem *cpu_boot_reg(int cpu)
|
|
|
|
{
|
|
|
|
void __iomem *boot_reg;
|
|
|
|
|
|
|
|
boot_reg = cpu_boot_reg_base();
|
|
|
|
if (soc_is_exynos4412())
|
|
|
|
boot_reg += 4*cpu;
|
2013-06-18 15:29:35 +00:00
|
|
|
else if (soc_is_exynos5420())
|
|
|
|
boot_reg += 4;
|
2012-11-24 02:13:48 +00:00
|
|
|
return boot_reg;
|
|
|
|
}
|
2011-07-16 04:39:09 +00:00
|
|
|
|
ARM: Fix subtle race in CPU pen_release hotplug code
There is a subtle race in the CPU hotplug code, where a CPU which has
been offlined can online itself before being requested, which results
in things going astray on the next online/offline cycle.
What happens in the normal online/offline/online cycle is:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads -1
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
However, as the write of -1 of pen_release is not fully flushed back to
memory, and the checking of pen_release is done with caches disabled,
this allows CPU3 the opportunity to read the old value of pen_release:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads 3
starts boot
pen_release = -1
requests boot of CPU3
pen_release = 3
flush cache line
Fix this by grouping the write of pen_release along with its cache line
flushing code to ensure that any update to pen_release is always pushed
out to physical memory.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2010-12-18 10:53:12 +00:00
|
|
|
/*
|
|
|
|
* Write pen_release in a way that is guaranteed to be visible to all
|
|
|
|
* observers, irrespective of whether they're taking part in coherency
|
|
|
|
* or not. This is necessary for the hotplug code to work reliably.
|
|
|
|
*/
|
|
|
|
static void write_pen_release(int val)
|
|
|
|
{
|
|
|
|
pen_release = val;
|
|
|
|
smp_wmb();
|
2013-12-05 19:26:16 +00:00
|
|
|
sync_cache_w(&pen_release);
|
ARM: Fix subtle race in CPU pen_release hotplug code
There is a subtle race in the CPU hotplug code, where a CPU which has
been offlined can online itself before being requested, which results
in things going astray on the next online/offline cycle.
What happens in the normal online/offline/online cycle is:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads -1
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
However, as the write of -1 of pen_release is not fully flushed back to
memory, and the checking of pen_release is done with caches disabled,
this allows CPU3 the opportunity to read the old value of pen_release:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads 3
starts boot
pen_release = -1
requests boot of CPU3
pen_release = 3
flush cache line
Fix this by grouping the write of pen_release along with its cache line
flushing code to ensure that any update to pen_release is always pushed
out to physical memory.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2010-12-18 10:53:12 +00:00
|
|
|
}
|
|
|
|
|
2010-07-26 12:08:52 +00:00
|
|
|
static void __iomem *scu_base_addr(void)
|
|
|
|
{
|
|
|
|
return (void __iomem *)(S5P_VA_SCU);
|
|
|
|
}
|
|
|
|
|
|
|
|
static DEFINE_SPINLOCK(boot_lock);
|
|
|
|
|
2013-06-17 19:43:14 +00:00
|
|
|
static void exynos_secondary_init(unsigned int cpu)
|
2010-07-26 12:08:52 +00:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
* let the primary processor know we're out of the
|
|
|
|
* pen, then head off into the C entry point
|
|
|
|
*/
|
ARM: Fix subtle race in CPU pen_release hotplug code
There is a subtle race in the CPU hotplug code, where a CPU which has
been offlined can online itself before being requested, which results
in things going astray on the next online/offline cycle.
What happens in the normal online/offline/online cycle is:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads -1
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
However, as the write of -1 of pen_release is not fully flushed back to
memory, and the checking of pen_release is done with caches disabled,
this allows CPU3 the opportunity to read the old value of pen_release:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads 3
starts boot
pen_release = -1
requests boot of CPU3
pen_release = 3
flush cache line
Fix this by grouping the write of pen_release along with its cache line
flushing code to ensure that any update to pen_release is always pushed
out to physical memory.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2010-12-18 10:53:12 +00:00
|
|
|
write_pen_release(-1);
|
2010-07-26 12:08:52 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Synchronise with the boot thread.
|
|
|
|
*/
|
|
|
|
spin_lock(&boot_lock);
|
|
|
|
spin_unlock(&boot_lock);
|
|
|
|
}
|
|
|
|
|
2013-06-17 19:43:14 +00:00
|
|
|
static int exynos_boot_secondary(unsigned int cpu, struct task_struct *idle)
|
2010-07-26 12:08:52 +00:00
|
|
|
{
|
|
|
|
unsigned long timeout;
|
2012-11-24 02:13:48 +00:00
|
|
|
unsigned long phys_cpu = cpu_logical_map(cpu);
|
2010-07-26 12:08:52 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Set synchronisation state between this boot processor
|
|
|
|
* and the secondary one
|
|
|
|
*/
|
|
|
|
spin_lock(&boot_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The secondary processor is waiting to be released from
|
|
|
|
* the holding pen - release it, then wait for it to flag
|
|
|
|
* that it has been released by resetting pen_release.
|
|
|
|
*
|
|
|
|
* Note that "pen_release" is the hardware CPU ID, whereas
|
|
|
|
* "cpu" is Linux's internal ID.
|
|
|
|
*/
|
2012-11-24 02:13:48 +00:00
|
|
|
write_pen_release(phys_cpu);
|
2010-07-26 12:08:52 +00:00
|
|
|
|
2011-07-16 04:39:09 +00:00
|
|
|
if (!(__raw_readl(S5P_ARM_CORE1_STATUS) & S5P_CORE_LOCAL_PWR_EN)) {
|
|
|
|
__raw_writel(S5P_CORE_LOCAL_PWR_EN,
|
|
|
|
S5P_ARM_CORE1_CONFIGURATION);
|
|
|
|
|
|
|
|
timeout = 10;
|
|
|
|
|
|
|
|
/* wait max 10 ms until cpu1 is on */
|
|
|
|
while ((__raw_readl(S5P_ARM_CORE1_STATUS)
|
|
|
|
& S5P_CORE_LOCAL_PWR_EN) != S5P_CORE_LOCAL_PWR_EN) {
|
|
|
|
if (timeout-- == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
mdelay(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (timeout == 0) {
|
|
|
|
printk(KERN_ERR "cpu1 power enable failed");
|
|
|
|
spin_unlock(&boot_lock);
|
|
|
|
return -ETIMEDOUT;
|
|
|
|
}
|
|
|
|
}
|
2010-07-26 12:08:52 +00:00
|
|
|
/*
|
|
|
|
* Send the secondary CPU a soft interrupt, thereby causing
|
|
|
|
* the boot monitor to read the system wide flags register,
|
|
|
|
* and branch to the address found there.
|
|
|
|
*/
|
|
|
|
|
|
|
|
timeout = jiffies + (1 * HZ);
|
|
|
|
while (time_before(jiffies, timeout)) {
|
2012-12-11 04:58:43 +00:00
|
|
|
unsigned long boot_addr;
|
|
|
|
|
2010-07-26 12:08:52 +00:00
|
|
|
smp_rmb();
|
2011-07-16 04:39:09 +00:00
|
|
|
|
2012-12-11 04:58:43 +00:00
|
|
|
boot_addr = virt_to_phys(exynos4_secondary_startup);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Try to set boot address using firmware first
|
|
|
|
* and fall back to boot register if it fails.
|
|
|
|
*/
|
|
|
|
if (call_firmware_op(set_cpu_boot_addr, phys_cpu, boot_addr))
|
|
|
|
__raw_writel(boot_addr, cpu_boot_reg(phys_cpu));
|
|
|
|
|
|
|
|
call_firmware_op(cpu_boot, phys_cpu);
|
|
|
|
|
2012-11-26 21:05:48 +00:00
|
|
|
arch_send_wakeup_ipi_mask(cpumask_of(cpu));
|
2011-07-16 04:39:09 +00:00
|
|
|
|
2010-07-26 12:08:52 +00:00
|
|
|
if (pen_release == -1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
udelay(10);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* now the secondary core is starting up let it run its
|
|
|
|
* calibrations, then wait for it to finish
|
|
|
|
*/
|
|
|
|
spin_unlock(&boot_lock);
|
|
|
|
|
|
|
|
return pen_release != -1 ? -ENOSYS : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialise the CPU possible map early - this describes the CPUs
|
|
|
|
* which may be present or become present in the system.
|
|
|
|
*/
|
|
|
|
|
2011-09-08 12:15:22 +00:00
|
|
|
static void __init exynos_smp_init_cpus(void)
|
2010-07-26 12:08:52 +00:00
|
|
|
{
|
|
|
|
void __iomem *scu_base = scu_base_addr();
|
|
|
|
unsigned int i, ncores;
|
|
|
|
|
2013-06-18 15:29:34 +00:00
|
|
|
if (read_cpuid_part_number() == ARM_CPU_PART_CORTEX_A9)
|
2012-01-25 06:35:57 +00:00
|
|
|
ncores = scu_base ? scu_get_core_count(scu_base) : 1;
|
2013-06-18 15:29:34 +00:00
|
|
|
else
|
|
|
|
/*
|
|
|
|
* CPU Nodes are passed thru DT and set_cpu_possible
|
|
|
|
* is set by "arm_dt_init_cpu_maps".
|
|
|
|
*/
|
|
|
|
return;
|
2010-07-26 12:08:52 +00:00
|
|
|
|
|
|
|
/* sanity check */
|
2011-10-20 21:04:18 +00:00
|
|
|
if (ncores > nr_cpu_ids) {
|
|
|
|
pr_warn("SMP: %u cores greater than maximum (%u), clipping\n",
|
|
|
|
ncores, nr_cpu_ids);
|
|
|
|
ncores = nr_cpu_ids;
|
2010-07-26 12:08:52 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < ncores; i++)
|
|
|
|
set_cpu_possible(i, true);
|
|
|
|
}
|
|
|
|
|
2011-09-08 12:15:22 +00:00
|
|
|
static void __init exynos_smp_prepare_cpus(unsigned int max_cpus)
|
2010-07-26 12:08:52 +00:00
|
|
|
{
|
2012-11-24 02:13:48 +00:00
|
|
|
int i;
|
|
|
|
|
2013-06-10 09:28:04 +00:00
|
|
|
if (read_cpuid_part_number() == ARM_CPU_PART_CORTEX_A9)
|
2012-01-25 06:35:57 +00:00
|
|
|
scu_enable(scu_base_addr());
|
2010-12-03 11:09:48 +00:00
|
|
|
|
2010-07-26 12:08:52 +00:00
|
|
|
/*
|
2010-12-03 11:09:48 +00:00
|
|
|
* Write the address of secondary startup into the
|
|
|
|
* system-wide flags register. The boot monitor waits
|
|
|
|
* until it receives a soft interrupt, and then the
|
|
|
|
* secondary CPU branches to this address.
|
2012-12-11 04:58:43 +00:00
|
|
|
*
|
|
|
|
* Try using firmware operation first and fall back to
|
|
|
|
* boot register if it fails.
|
2010-07-26 12:08:52 +00:00
|
|
|
*/
|
2012-12-11 04:58:43 +00:00
|
|
|
for (i = 1; i < max_cpus; ++i) {
|
|
|
|
unsigned long phys_cpu;
|
|
|
|
unsigned long boot_addr;
|
|
|
|
|
|
|
|
phys_cpu = cpu_logical_map(i);
|
|
|
|
boot_addr = virt_to_phys(exynos4_secondary_startup);
|
|
|
|
|
|
|
|
if (call_firmware_op(set_cpu_boot_addr, phys_cpu, boot_addr))
|
|
|
|
__raw_writel(boot_addr, cpu_boot_reg(phys_cpu));
|
|
|
|
}
|
2010-07-26 12:08:52 +00:00
|
|
|
}
|
2011-09-08 12:15:22 +00:00
|
|
|
|
|
|
|
struct smp_operations exynos_smp_ops __initdata = {
|
|
|
|
.smp_init_cpus = exynos_smp_init_cpus,
|
|
|
|
.smp_prepare_cpus = exynos_smp_prepare_cpus,
|
|
|
|
.smp_secondary_init = exynos_secondary_init,
|
|
|
|
.smp_boot_secondary = exynos_boot_secondary,
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
|
.cpu_die = exynos_cpu_die,
|
|
|
|
#endif
|
|
|
|
};
|