mirror of
https://github.com/torvalds/linux.git
synced 2024-12-21 10:31:54 +00:00
69111bac42
This still has not been merged and now powerpc is the only arch that does not have this change. Sorry about missing linuxppc-dev before. V2->V2 - Fix up to work against 3.18-rc1 __get_cpu_var() is used for multiple purposes in the kernel source. One of them is address calculation via the form &__get_cpu_var(x). This calculates the address for the instance of the percpu variable of the current processor based on an offset. Other use cases are for storing and retrieving data from the current processors percpu area. __get_cpu_var() can be used as an lvalue when writing data or on the right side of an assignment. __get_cpu_var() is defined as : __get_cpu_var() always only does an address determination. However, store and retrieve operations could use a segment prefix (or global register on other platforms) to avoid the address calculation. this_cpu_write() and this_cpu_read() can directly take an offset into a percpu area and use optimized assembly code to read and write per cpu variables. This patch converts __get_cpu_var into either an explicit address calculation using this_cpu_ptr() or into a use of this_cpu operations that use the offset. Thereby address calculations are avoided and less registers are used when code is generated. At the end of the patch set all uses of __get_cpu_var have been removed so the macro is removed too. The patch set includes passes over all arches as well. Once these operations are used throughout then specialized macros can be defined in non -x86 arches as well in order to optimize per cpu access by f.e. using a global register that may be set to the per cpu base. Transformations done to __get_cpu_var() 1. Determine the address of the percpu instance of the current processor. DEFINE_PER_CPU(int, y); int *x = &__get_cpu_var(y); Converts to int *x = this_cpu_ptr(&y); 2. Same as #1 but this time an array structure is involved. DEFINE_PER_CPU(int, y[20]); int *x = __get_cpu_var(y); Converts to int *x = this_cpu_ptr(y); 3. Retrieve the content of the current processors instance of a per cpu variable. DEFINE_PER_CPU(int, y); int x = __get_cpu_var(y) Converts to int x = __this_cpu_read(y); 4. Retrieve the content of a percpu struct DEFINE_PER_CPU(struct mystruct, y); struct mystruct x = __get_cpu_var(y); Converts to memcpy(&x, this_cpu_ptr(&y), sizeof(x)); 5. Assignment to a per cpu variable DEFINE_PER_CPU(int, y) __get_cpu_var(y) = x; Converts to __this_cpu_write(y, x); 6. Increment/Decrement etc of a per cpu variable DEFINE_PER_CPU(int, y); __get_cpu_var(y)++ Converts to __this_cpu_inc(y) Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> CC: Paul Mackerras <paulus@samba.org> Signed-off-by: Christoph Lameter <cl@linux.com> [mpe: Fix build errors caused by set/or_softirq_pending(), and rework assignment in __set_breakpoint() to use memcpy().] Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
1028 lines
24 KiB
C
1028 lines
24 KiB
C
#include <linux/device.h>
|
|
#include <linux/cpu.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/init.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/export.h>
|
|
#include <linux/nodemask.h>
|
|
#include <linux/cpumask.h>
|
|
#include <linux/notifier.h>
|
|
|
|
#include <asm/current.h>
|
|
#include <asm/processor.h>
|
|
#include <asm/cputable.h>
|
|
#include <asm/hvcall.h>
|
|
#include <asm/prom.h>
|
|
#include <asm/machdep.h>
|
|
#include <asm/smp.h>
|
|
#include <asm/pmc.h>
|
|
#include <asm/firmware.h>
|
|
|
|
#include "cacheinfo.h"
|
|
|
|
#ifdef CONFIG_PPC64
|
|
#include <asm/paca.h>
|
|
#include <asm/lppaca.h>
|
|
#endif
|
|
|
|
static DEFINE_PER_CPU(struct cpu, cpu_devices);
|
|
|
|
/*
|
|
* SMT snooze delay stuff, 64-bit only for now
|
|
*/
|
|
|
|
#ifdef CONFIG_PPC64
|
|
|
|
/* Time in microseconds we delay before sleeping in the idle loop */
|
|
DEFINE_PER_CPU(long, smt_snooze_delay) = { 100 };
|
|
|
|
static ssize_t store_smt_snooze_delay(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
struct cpu *cpu = container_of(dev, struct cpu, dev);
|
|
ssize_t ret;
|
|
long snooze;
|
|
|
|
ret = sscanf(buf, "%ld", &snooze);
|
|
if (ret != 1)
|
|
return -EINVAL;
|
|
|
|
per_cpu(smt_snooze_delay, cpu->dev.id) = snooze;
|
|
return count;
|
|
}
|
|
|
|
static ssize_t show_smt_snooze_delay(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct cpu *cpu = container_of(dev, struct cpu, dev);
|
|
|
|
return sprintf(buf, "%ld\n", per_cpu(smt_snooze_delay, cpu->dev.id));
|
|
}
|
|
|
|
static DEVICE_ATTR(smt_snooze_delay, 0644, show_smt_snooze_delay,
|
|
store_smt_snooze_delay);
|
|
|
|
static int __init setup_smt_snooze_delay(char *str)
|
|
{
|
|
unsigned int cpu;
|
|
long snooze;
|
|
|
|
if (!cpu_has_feature(CPU_FTR_SMT))
|
|
return 1;
|
|
|
|
snooze = simple_strtol(str, NULL, 10);
|
|
for_each_possible_cpu(cpu)
|
|
per_cpu(smt_snooze_delay, cpu) = snooze;
|
|
|
|
return 1;
|
|
}
|
|
__setup("smt-snooze-delay=", setup_smt_snooze_delay);
|
|
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#ifdef CONFIG_PPC_FSL_BOOK3E
|
|
#define MAX_BIT 63
|
|
|
|
static u64 pw20_wt;
|
|
static u64 altivec_idle_wt;
|
|
|
|
static unsigned int get_idle_ticks_bit(u64 ns)
|
|
{
|
|
u64 cycle;
|
|
|
|
if (ns >= 10000)
|
|
cycle = div_u64(ns + 500, 1000) * tb_ticks_per_usec;
|
|
else
|
|
cycle = div_u64(ns * tb_ticks_per_usec, 1000);
|
|
|
|
if (!cycle)
|
|
return 0;
|
|
|
|
return ilog2(cycle);
|
|
}
|
|
|
|
static void do_show_pwrmgtcr0(void *val)
|
|
{
|
|
u32 *value = val;
|
|
|
|
*value = mfspr(SPRN_PWRMGTCR0);
|
|
}
|
|
|
|
static ssize_t show_pw20_state(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
u32 value;
|
|
unsigned int cpu = dev->id;
|
|
|
|
smp_call_function_single(cpu, do_show_pwrmgtcr0, &value, 1);
|
|
|
|
value &= PWRMGTCR0_PW20_WAIT;
|
|
|
|
return sprintf(buf, "%u\n", value ? 1 : 0);
|
|
}
|
|
|
|
static void do_store_pw20_state(void *val)
|
|
{
|
|
u32 *value = val;
|
|
u32 pw20_state;
|
|
|
|
pw20_state = mfspr(SPRN_PWRMGTCR0);
|
|
|
|
if (*value)
|
|
pw20_state |= PWRMGTCR0_PW20_WAIT;
|
|
else
|
|
pw20_state &= ~PWRMGTCR0_PW20_WAIT;
|
|
|
|
mtspr(SPRN_PWRMGTCR0, pw20_state);
|
|
}
|
|
|
|
static ssize_t store_pw20_state(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
u32 value;
|
|
unsigned int cpu = dev->id;
|
|
|
|
if (kstrtou32(buf, 0, &value))
|
|
return -EINVAL;
|
|
|
|
if (value > 1)
|
|
return -EINVAL;
|
|
|
|
smp_call_function_single(cpu, do_store_pw20_state, &value, 1);
|
|
|
|
return count;
|
|
}
|
|
|
|
static ssize_t show_pw20_wait_time(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
u32 value;
|
|
u64 tb_cycle = 1;
|
|
u64 time;
|
|
|
|
unsigned int cpu = dev->id;
|
|
|
|
if (!pw20_wt) {
|
|
smp_call_function_single(cpu, do_show_pwrmgtcr0, &value, 1);
|
|
value = (value & PWRMGTCR0_PW20_ENT) >>
|
|
PWRMGTCR0_PW20_ENT_SHIFT;
|
|
|
|
tb_cycle = (tb_cycle << (MAX_BIT - value + 1));
|
|
/* convert ms to ns */
|
|
if (tb_ticks_per_usec > 1000) {
|
|
time = div_u64(tb_cycle, tb_ticks_per_usec / 1000);
|
|
} else {
|
|
u32 rem_us;
|
|
|
|
time = div_u64_rem(tb_cycle, tb_ticks_per_usec,
|
|
&rem_us);
|
|
time = time * 1000 + rem_us * 1000 / tb_ticks_per_usec;
|
|
}
|
|
} else {
|
|
time = pw20_wt;
|
|
}
|
|
|
|
return sprintf(buf, "%llu\n", time > 0 ? time : 0);
|
|
}
|
|
|
|
static void set_pw20_wait_entry_bit(void *val)
|
|
{
|
|
u32 *value = val;
|
|
u32 pw20_idle;
|
|
|
|
pw20_idle = mfspr(SPRN_PWRMGTCR0);
|
|
|
|
/* Set Automatic PW20 Core Idle Count */
|
|
/* clear count */
|
|
pw20_idle &= ~PWRMGTCR0_PW20_ENT;
|
|
|
|
/* set count */
|
|
pw20_idle |= ((MAX_BIT - *value) << PWRMGTCR0_PW20_ENT_SHIFT);
|
|
|
|
mtspr(SPRN_PWRMGTCR0, pw20_idle);
|
|
}
|
|
|
|
static ssize_t store_pw20_wait_time(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
u32 entry_bit;
|
|
u64 value;
|
|
|
|
unsigned int cpu = dev->id;
|
|
|
|
if (kstrtou64(buf, 0, &value))
|
|
return -EINVAL;
|
|
|
|
if (!value)
|
|
return -EINVAL;
|
|
|
|
entry_bit = get_idle_ticks_bit(value);
|
|
if (entry_bit > MAX_BIT)
|
|
return -EINVAL;
|
|
|
|
pw20_wt = value;
|
|
|
|
smp_call_function_single(cpu, set_pw20_wait_entry_bit,
|
|
&entry_bit, 1);
|
|
|
|
return count;
|
|
}
|
|
|
|
static ssize_t show_altivec_idle(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
u32 value;
|
|
unsigned int cpu = dev->id;
|
|
|
|
smp_call_function_single(cpu, do_show_pwrmgtcr0, &value, 1);
|
|
|
|
value &= PWRMGTCR0_AV_IDLE_PD_EN;
|
|
|
|
return sprintf(buf, "%u\n", value ? 1 : 0);
|
|
}
|
|
|
|
static void do_store_altivec_idle(void *val)
|
|
{
|
|
u32 *value = val;
|
|
u32 altivec_idle;
|
|
|
|
altivec_idle = mfspr(SPRN_PWRMGTCR0);
|
|
|
|
if (*value)
|
|
altivec_idle |= PWRMGTCR0_AV_IDLE_PD_EN;
|
|
else
|
|
altivec_idle &= ~PWRMGTCR0_AV_IDLE_PD_EN;
|
|
|
|
mtspr(SPRN_PWRMGTCR0, altivec_idle);
|
|
}
|
|
|
|
static ssize_t store_altivec_idle(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
u32 value;
|
|
unsigned int cpu = dev->id;
|
|
|
|
if (kstrtou32(buf, 0, &value))
|
|
return -EINVAL;
|
|
|
|
if (value > 1)
|
|
return -EINVAL;
|
|
|
|
smp_call_function_single(cpu, do_store_altivec_idle, &value, 1);
|
|
|
|
return count;
|
|
}
|
|
|
|
static ssize_t show_altivec_idle_wait_time(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
u32 value;
|
|
u64 tb_cycle = 1;
|
|
u64 time;
|
|
|
|
unsigned int cpu = dev->id;
|
|
|
|
if (!altivec_idle_wt) {
|
|
smp_call_function_single(cpu, do_show_pwrmgtcr0, &value, 1);
|
|
value = (value & PWRMGTCR0_AV_IDLE_CNT) >>
|
|
PWRMGTCR0_AV_IDLE_CNT_SHIFT;
|
|
|
|
tb_cycle = (tb_cycle << (MAX_BIT - value + 1));
|
|
/* convert ms to ns */
|
|
if (tb_ticks_per_usec > 1000) {
|
|
time = div_u64(tb_cycle, tb_ticks_per_usec / 1000);
|
|
} else {
|
|
u32 rem_us;
|
|
|
|
time = div_u64_rem(tb_cycle, tb_ticks_per_usec,
|
|
&rem_us);
|
|
time = time * 1000 + rem_us * 1000 / tb_ticks_per_usec;
|
|
}
|
|
} else {
|
|
time = altivec_idle_wt;
|
|
}
|
|
|
|
return sprintf(buf, "%llu\n", time > 0 ? time : 0);
|
|
}
|
|
|
|
static void set_altivec_idle_wait_entry_bit(void *val)
|
|
{
|
|
u32 *value = val;
|
|
u32 altivec_idle;
|
|
|
|
altivec_idle = mfspr(SPRN_PWRMGTCR0);
|
|
|
|
/* Set Automatic AltiVec Idle Count */
|
|
/* clear count */
|
|
altivec_idle &= ~PWRMGTCR0_AV_IDLE_CNT;
|
|
|
|
/* set count */
|
|
altivec_idle |= ((MAX_BIT - *value) << PWRMGTCR0_AV_IDLE_CNT_SHIFT);
|
|
|
|
mtspr(SPRN_PWRMGTCR0, altivec_idle);
|
|
}
|
|
|
|
static ssize_t store_altivec_idle_wait_time(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
u32 entry_bit;
|
|
u64 value;
|
|
|
|
unsigned int cpu = dev->id;
|
|
|
|
if (kstrtou64(buf, 0, &value))
|
|
return -EINVAL;
|
|
|
|
if (!value)
|
|
return -EINVAL;
|
|
|
|
entry_bit = get_idle_ticks_bit(value);
|
|
if (entry_bit > MAX_BIT)
|
|
return -EINVAL;
|
|
|
|
altivec_idle_wt = value;
|
|
|
|
smp_call_function_single(cpu, set_altivec_idle_wait_entry_bit,
|
|
&entry_bit, 1);
|
|
|
|
return count;
|
|
}
|
|
|
|
/*
|
|
* Enable/Disable interface:
|
|
* 0, disable. 1, enable.
|
|
*/
|
|
static DEVICE_ATTR(pw20_state, 0600, show_pw20_state, store_pw20_state);
|
|
static DEVICE_ATTR(altivec_idle, 0600, show_altivec_idle, store_altivec_idle);
|
|
|
|
/*
|
|
* Set wait time interface:(Nanosecond)
|
|
* Example: Base on TBfreq is 41MHZ.
|
|
* 1~48(ns): TB[63]
|
|
* 49~97(ns): TB[62]
|
|
* 98~195(ns): TB[61]
|
|
* 196~390(ns): TB[60]
|
|
* 391~780(ns): TB[59]
|
|
* 781~1560(ns): TB[58]
|
|
* ...
|
|
*/
|
|
static DEVICE_ATTR(pw20_wait_time, 0600,
|
|
show_pw20_wait_time,
|
|
store_pw20_wait_time);
|
|
static DEVICE_ATTR(altivec_idle_wait_time, 0600,
|
|
show_altivec_idle_wait_time,
|
|
store_altivec_idle_wait_time);
|
|
#endif
|
|
|
|
/*
|
|
* Enabling PMCs will slow partition context switch times so we only do
|
|
* it the first time we write to the PMCs.
|
|
*/
|
|
|
|
static DEFINE_PER_CPU(char, pmcs_enabled);
|
|
|
|
void ppc_enable_pmcs(void)
|
|
{
|
|
ppc_set_pmu_inuse(1);
|
|
|
|
/* Only need to enable them once */
|
|
if (__this_cpu_read(pmcs_enabled))
|
|
return;
|
|
|
|
__this_cpu_write(pmcs_enabled, 1);
|
|
|
|
if (ppc_md.enable_pmcs)
|
|
ppc_md.enable_pmcs();
|
|
}
|
|
EXPORT_SYMBOL(ppc_enable_pmcs);
|
|
|
|
#define __SYSFS_SPRSETUP_READ_WRITE(NAME, ADDRESS, EXTRA) \
|
|
static void read_##NAME(void *val) \
|
|
{ \
|
|
*(unsigned long *)val = mfspr(ADDRESS); \
|
|
} \
|
|
static void write_##NAME(void *val) \
|
|
{ \
|
|
EXTRA; \
|
|
mtspr(ADDRESS, *(unsigned long *)val); \
|
|
}
|
|
|
|
#define __SYSFS_SPRSETUP_SHOW_STORE(NAME) \
|
|
static ssize_t show_##NAME(struct device *dev, \
|
|
struct device_attribute *attr, \
|
|
char *buf) \
|
|
{ \
|
|
struct cpu *cpu = container_of(dev, struct cpu, dev); \
|
|
unsigned long val; \
|
|
smp_call_function_single(cpu->dev.id, read_##NAME, &val, 1); \
|
|
return sprintf(buf, "%lx\n", val); \
|
|
} \
|
|
static ssize_t __used \
|
|
store_##NAME(struct device *dev, struct device_attribute *attr, \
|
|
const char *buf, size_t count) \
|
|
{ \
|
|
struct cpu *cpu = container_of(dev, struct cpu, dev); \
|
|
unsigned long val; \
|
|
int ret = sscanf(buf, "%lx", &val); \
|
|
if (ret != 1) \
|
|
return -EINVAL; \
|
|
smp_call_function_single(cpu->dev.id, write_##NAME, &val, 1); \
|
|
return count; \
|
|
}
|
|
|
|
#define SYSFS_PMCSETUP(NAME, ADDRESS) \
|
|
__SYSFS_SPRSETUP_READ_WRITE(NAME, ADDRESS, ppc_enable_pmcs()) \
|
|
__SYSFS_SPRSETUP_SHOW_STORE(NAME)
|
|
#define SYSFS_SPRSETUP(NAME, ADDRESS) \
|
|
__SYSFS_SPRSETUP_READ_WRITE(NAME, ADDRESS, ) \
|
|
__SYSFS_SPRSETUP_SHOW_STORE(NAME)
|
|
|
|
#define SYSFS_SPRSETUP_SHOW_STORE(NAME) \
|
|
__SYSFS_SPRSETUP_SHOW_STORE(NAME)
|
|
|
|
/* Let's define all possible registers, we'll only hook up the ones
|
|
* that are implemented on the current processor
|
|
*/
|
|
|
|
#if defined(CONFIG_PPC64)
|
|
#define HAS_PPC_PMC_CLASSIC 1
|
|
#define HAS_PPC_PMC_IBM 1
|
|
#define HAS_PPC_PMC_PA6T 1
|
|
#elif defined(CONFIG_6xx)
|
|
#define HAS_PPC_PMC_CLASSIC 1
|
|
#define HAS_PPC_PMC_IBM 1
|
|
#define HAS_PPC_PMC_G4 1
|
|
#endif
|
|
|
|
|
|
#ifdef HAS_PPC_PMC_CLASSIC
|
|
SYSFS_PMCSETUP(mmcr0, SPRN_MMCR0);
|
|
SYSFS_PMCSETUP(mmcr1, SPRN_MMCR1);
|
|
SYSFS_PMCSETUP(pmc1, SPRN_PMC1);
|
|
SYSFS_PMCSETUP(pmc2, SPRN_PMC2);
|
|
SYSFS_PMCSETUP(pmc3, SPRN_PMC3);
|
|
SYSFS_PMCSETUP(pmc4, SPRN_PMC4);
|
|
SYSFS_PMCSETUP(pmc5, SPRN_PMC5);
|
|
SYSFS_PMCSETUP(pmc6, SPRN_PMC6);
|
|
|
|
#ifdef HAS_PPC_PMC_G4
|
|
SYSFS_PMCSETUP(mmcr2, SPRN_MMCR2);
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC64
|
|
SYSFS_PMCSETUP(pmc7, SPRN_PMC7);
|
|
SYSFS_PMCSETUP(pmc8, SPRN_PMC8);
|
|
|
|
SYSFS_PMCSETUP(mmcra, SPRN_MMCRA);
|
|
SYSFS_SPRSETUP(purr, SPRN_PURR);
|
|
SYSFS_SPRSETUP(spurr, SPRN_SPURR);
|
|
SYSFS_SPRSETUP(pir, SPRN_PIR);
|
|
|
|
/*
|
|
Lets only enable read for phyp resources and
|
|
enable write when needed with a separate function.
|
|
Lets be conservative and default to pseries.
|
|
*/
|
|
static DEVICE_ATTR(mmcra, 0600, show_mmcra, store_mmcra);
|
|
static DEVICE_ATTR(spurr, 0400, show_spurr, NULL);
|
|
static DEVICE_ATTR(purr, 0400, show_purr, store_purr);
|
|
static DEVICE_ATTR(pir, 0400, show_pir, NULL);
|
|
|
|
static unsigned long dscr_default;
|
|
|
|
static void read_dscr(void *val)
|
|
{
|
|
*(unsigned long *)val = get_paca()->dscr_default;
|
|
}
|
|
|
|
static void write_dscr(void *val)
|
|
{
|
|
get_paca()->dscr_default = *(unsigned long *)val;
|
|
if (!current->thread.dscr_inherit) {
|
|
current->thread.dscr = *(unsigned long *)val;
|
|
mtspr(SPRN_DSCR, *(unsigned long *)val);
|
|
}
|
|
}
|
|
|
|
SYSFS_SPRSETUP_SHOW_STORE(dscr);
|
|
static DEVICE_ATTR(dscr, 0600, show_dscr, store_dscr);
|
|
|
|
static void add_write_permission_dev_attr(struct device_attribute *attr)
|
|
{
|
|
attr->attr.mode |= 0200;
|
|
}
|
|
|
|
static ssize_t show_dscr_default(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
return sprintf(buf, "%lx\n", dscr_default);
|
|
}
|
|
|
|
static ssize_t __used store_dscr_default(struct device *dev,
|
|
struct device_attribute *attr, const char *buf,
|
|
size_t count)
|
|
{
|
|
unsigned long val;
|
|
int ret = 0;
|
|
|
|
ret = sscanf(buf, "%lx", &val);
|
|
if (ret != 1)
|
|
return -EINVAL;
|
|
dscr_default = val;
|
|
|
|
on_each_cpu(write_dscr, &val, 1);
|
|
|
|
return count;
|
|
}
|
|
|
|
static DEVICE_ATTR(dscr_default, 0600,
|
|
show_dscr_default, store_dscr_default);
|
|
|
|
static void sysfs_create_dscr_default(void)
|
|
{
|
|
int err = 0;
|
|
if (cpu_has_feature(CPU_FTR_DSCR))
|
|
err = device_create_file(cpu_subsys.dev_root, &dev_attr_dscr_default);
|
|
}
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#ifdef HAS_PPC_PMC_PA6T
|
|
SYSFS_PMCSETUP(pa6t_pmc0, SPRN_PA6T_PMC0);
|
|
SYSFS_PMCSETUP(pa6t_pmc1, SPRN_PA6T_PMC1);
|
|
SYSFS_PMCSETUP(pa6t_pmc2, SPRN_PA6T_PMC2);
|
|
SYSFS_PMCSETUP(pa6t_pmc3, SPRN_PA6T_PMC3);
|
|
SYSFS_PMCSETUP(pa6t_pmc4, SPRN_PA6T_PMC4);
|
|
SYSFS_PMCSETUP(pa6t_pmc5, SPRN_PA6T_PMC5);
|
|
#ifdef CONFIG_DEBUG_KERNEL
|
|
SYSFS_SPRSETUP(hid0, SPRN_HID0);
|
|
SYSFS_SPRSETUP(hid1, SPRN_HID1);
|
|
SYSFS_SPRSETUP(hid4, SPRN_HID4);
|
|
SYSFS_SPRSETUP(hid5, SPRN_HID5);
|
|
SYSFS_SPRSETUP(ima0, SPRN_PA6T_IMA0);
|
|
SYSFS_SPRSETUP(ima1, SPRN_PA6T_IMA1);
|
|
SYSFS_SPRSETUP(ima2, SPRN_PA6T_IMA2);
|
|
SYSFS_SPRSETUP(ima3, SPRN_PA6T_IMA3);
|
|
SYSFS_SPRSETUP(ima4, SPRN_PA6T_IMA4);
|
|
SYSFS_SPRSETUP(ima5, SPRN_PA6T_IMA5);
|
|
SYSFS_SPRSETUP(ima6, SPRN_PA6T_IMA6);
|
|
SYSFS_SPRSETUP(ima7, SPRN_PA6T_IMA7);
|
|
SYSFS_SPRSETUP(ima8, SPRN_PA6T_IMA8);
|
|
SYSFS_SPRSETUP(ima9, SPRN_PA6T_IMA9);
|
|
SYSFS_SPRSETUP(imaat, SPRN_PA6T_IMAAT);
|
|
SYSFS_SPRSETUP(btcr, SPRN_PA6T_BTCR);
|
|
SYSFS_SPRSETUP(pccr, SPRN_PA6T_PCCR);
|
|
SYSFS_SPRSETUP(rpccr, SPRN_PA6T_RPCCR);
|
|
SYSFS_SPRSETUP(der, SPRN_PA6T_DER);
|
|
SYSFS_SPRSETUP(mer, SPRN_PA6T_MER);
|
|
SYSFS_SPRSETUP(ber, SPRN_PA6T_BER);
|
|
SYSFS_SPRSETUP(ier, SPRN_PA6T_IER);
|
|
SYSFS_SPRSETUP(sier, SPRN_PA6T_SIER);
|
|
SYSFS_SPRSETUP(siar, SPRN_PA6T_SIAR);
|
|
SYSFS_SPRSETUP(tsr0, SPRN_PA6T_TSR0);
|
|
SYSFS_SPRSETUP(tsr1, SPRN_PA6T_TSR1);
|
|
SYSFS_SPRSETUP(tsr2, SPRN_PA6T_TSR2);
|
|
SYSFS_SPRSETUP(tsr3, SPRN_PA6T_TSR3);
|
|
#endif /* CONFIG_DEBUG_KERNEL */
|
|
#endif /* HAS_PPC_PMC_PA6T */
|
|
|
|
#ifdef HAS_PPC_PMC_IBM
|
|
static struct device_attribute ibm_common_attrs[] = {
|
|
__ATTR(mmcr0, 0600, show_mmcr0, store_mmcr0),
|
|
__ATTR(mmcr1, 0600, show_mmcr1, store_mmcr1),
|
|
};
|
|
#endif /* HAS_PPC_PMC_G4 */
|
|
|
|
#ifdef HAS_PPC_PMC_G4
|
|
static struct device_attribute g4_common_attrs[] = {
|
|
__ATTR(mmcr0, 0600, show_mmcr0, store_mmcr0),
|
|
__ATTR(mmcr1, 0600, show_mmcr1, store_mmcr1),
|
|
__ATTR(mmcr2, 0600, show_mmcr2, store_mmcr2),
|
|
};
|
|
#endif /* HAS_PPC_PMC_G4 */
|
|
|
|
static struct device_attribute classic_pmc_attrs[] = {
|
|
__ATTR(pmc1, 0600, show_pmc1, store_pmc1),
|
|
__ATTR(pmc2, 0600, show_pmc2, store_pmc2),
|
|
__ATTR(pmc3, 0600, show_pmc3, store_pmc3),
|
|
__ATTR(pmc4, 0600, show_pmc4, store_pmc4),
|
|
__ATTR(pmc5, 0600, show_pmc5, store_pmc5),
|
|
__ATTR(pmc6, 0600, show_pmc6, store_pmc6),
|
|
#ifdef CONFIG_PPC64
|
|
__ATTR(pmc7, 0600, show_pmc7, store_pmc7),
|
|
__ATTR(pmc8, 0600, show_pmc8, store_pmc8),
|
|
#endif
|
|
};
|
|
|
|
#ifdef HAS_PPC_PMC_PA6T
|
|
static struct device_attribute pa6t_attrs[] = {
|
|
__ATTR(mmcr0, 0600, show_mmcr0, store_mmcr0),
|
|
__ATTR(mmcr1, 0600, show_mmcr1, store_mmcr1),
|
|
__ATTR(pmc0, 0600, show_pa6t_pmc0, store_pa6t_pmc0),
|
|
__ATTR(pmc1, 0600, show_pa6t_pmc1, store_pa6t_pmc1),
|
|
__ATTR(pmc2, 0600, show_pa6t_pmc2, store_pa6t_pmc2),
|
|
__ATTR(pmc3, 0600, show_pa6t_pmc3, store_pa6t_pmc3),
|
|
__ATTR(pmc4, 0600, show_pa6t_pmc4, store_pa6t_pmc4),
|
|
__ATTR(pmc5, 0600, show_pa6t_pmc5, store_pa6t_pmc5),
|
|
#ifdef CONFIG_DEBUG_KERNEL
|
|
__ATTR(hid0, 0600, show_hid0, store_hid0),
|
|
__ATTR(hid1, 0600, show_hid1, store_hid1),
|
|
__ATTR(hid4, 0600, show_hid4, store_hid4),
|
|
__ATTR(hid5, 0600, show_hid5, store_hid5),
|
|
__ATTR(ima0, 0600, show_ima0, store_ima0),
|
|
__ATTR(ima1, 0600, show_ima1, store_ima1),
|
|
__ATTR(ima2, 0600, show_ima2, store_ima2),
|
|
__ATTR(ima3, 0600, show_ima3, store_ima3),
|
|
__ATTR(ima4, 0600, show_ima4, store_ima4),
|
|
__ATTR(ima5, 0600, show_ima5, store_ima5),
|
|
__ATTR(ima6, 0600, show_ima6, store_ima6),
|
|
__ATTR(ima7, 0600, show_ima7, store_ima7),
|
|
__ATTR(ima8, 0600, show_ima8, store_ima8),
|
|
__ATTR(ima9, 0600, show_ima9, store_ima9),
|
|
__ATTR(imaat, 0600, show_imaat, store_imaat),
|
|
__ATTR(btcr, 0600, show_btcr, store_btcr),
|
|
__ATTR(pccr, 0600, show_pccr, store_pccr),
|
|
__ATTR(rpccr, 0600, show_rpccr, store_rpccr),
|
|
__ATTR(der, 0600, show_der, store_der),
|
|
__ATTR(mer, 0600, show_mer, store_mer),
|
|
__ATTR(ber, 0600, show_ber, store_ber),
|
|
__ATTR(ier, 0600, show_ier, store_ier),
|
|
__ATTR(sier, 0600, show_sier, store_sier),
|
|
__ATTR(siar, 0600, show_siar, store_siar),
|
|
__ATTR(tsr0, 0600, show_tsr0, store_tsr0),
|
|
__ATTR(tsr1, 0600, show_tsr1, store_tsr1),
|
|
__ATTR(tsr2, 0600, show_tsr2, store_tsr2),
|
|
__ATTR(tsr3, 0600, show_tsr3, store_tsr3),
|
|
#endif /* CONFIG_DEBUG_KERNEL */
|
|
};
|
|
#endif /* HAS_PPC_PMC_PA6T */
|
|
#endif /* HAS_PPC_PMC_CLASSIC */
|
|
|
|
static void register_cpu_online(unsigned int cpu)
|
|
{
|
|
struct cpu *c = &per_cpu(cpu_devices, cpu);
|
|
struct device *s = &c->dev;
|
|
struct device_attribute *attrs, *pmc_attrs;
|
|
int i, nattrs;
|
|
|
|
#ifdef CONFIG_PPC64
|
|
if (cpu_has_feature(CPU_FTR_SMT))
|
|
device_create_file(s, &dev_attr_smt_snooze_delay);
|
|
#endif
|
|
|
|
/* PMC stuff */
|
|
switch (cur_cpu_spec->pmc_type) {
|
|
#ifdef HAS_PPC_PMC_IBM
|
|
case PPC_PMC_IBM:
|
|
attrs = ibm_common_attrs;
|
|
nattrs = sizeof(ibm_common_attrs) / sizeof(struct device_attribute);
|
|
pmc_attrs = classic_pmc_attrs;
|
|
break;
|
|
#endif /* HAS_PPC_PMC_IBM */
|
|
#ifdef HAS_PPC_PMC_G4
|
|
case PPC_PMC_G4:
|
|
attrs = g4_common_attrs;
|
|
nattrs = sizeof(g4_common_attrs) / sizeof(struct device_attribute);
|
|
pmc_attrs = classic_pmc_attrs;
|
|
break;
|
|
#endif /* HAS_PPC_PMC_G4 */
|
|
#ifdef HAS_PPC_PMC_PA6T
|
|
case PPC_PMC_PA6T:
|
|
/* PA Semi starts counting at PMC0 */
|
|
attrs = pa6t_attrs;
|
|
nattrs = sizeof(pa6t_attrs) / sizeof(struct device_attribute);
|
|
pmc_attrs = NULL;
|
|
break;
|
|
#endif /* HAS_PPC_PMC_PA6T */
|
|
default:
|
|
attrs = NULL;
|
|
nattrs = 0;
|
|
pmc_attrs = NULL;
|
|
}
|
|
|
|
for (i = 0; i < nattrs; i++)
|
|
device_create_file(s, &attrs[i]);
|
|
|
|
if (pmc_attrs)
|
|
for (i = 0; i < cur_cpu_spec->num_pmcs; i++)
|
|
device_create_file(s, &pmc_attrs[i]);
|
|
|
|
#ifdef CONFIG_PPC64
|
|
if (cpu_has_feature(CPU_FTR_MMCRA))
|
|
device_create_file(s, &dev_attr_mmcra);
|
|
|
|
if (cpu_has_feature(CPU_FTR_PURR)) {
|
|
if (!firmware_has_feature(FW_FEATURE_LPAR))
|
|
add_write_permission_dev_attr(&dev_attr_purr);
|
|
device_create_file(s, &dev_attr_purr);
|
|
}
|
|
|
|
if (cpu_has_feature(CPU_FTR_SPURR))
|
|
device_create_file(s, &dev_attr_spurr);
|
|
|
|
if (cpu_has_feature(CPU_FTR_DSCR))
|
|
device_create_file(s, &dev_attr_dscr);
|
|
|
|
if (cpu_has_feature(CPU_FTR_PPCAS_ARCH_V2))
|
|
device_create_file(s, &dev_attr_pir);
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#ifdef CONFIG_PPC_FSL_BOOK3E
|
|
if (PVR_VER(cur_cpu_spec->pvr_value) == PVR_VER_E6500) {
|
|
device_create_file(s, &dev_attr_pw20_state);
|
|
device_create_file(s, &dev_attr_pw20_wait_time);
|
|
|
|
device_create_file(s, &dev_attr_altivec_idle);
|
|
device_create_file(s, &dev_attr_altivec_idle_wait_time);
|
|
}
|
|
#endif
|
|
cacheinfo_cpu_online(cpu);
|
|
}
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
static void unregister_cpu_online(unsigned int cpu)
|
|
{
|
|
struct cpu *c = &per_cpu(cpu_devices, cpu);
|
|
struct device *s = &c->dev;
|
|
struct device_attribute *attrs, *pmc_attrs;
|
|
int i, nattrs;
|
|
|
|
BUG_ON(!c->hotpluggable);
|
|
|
|
#ifdef CONFIG_PPC64
|
|
if (cpu_has_feature(CPU_FTR_SMT))
|
|
device_remove_file(s, &dev_attr_smt_snooze_delay);
|
|
#endif
|
|
|
|
/* PMC stuff */
|
|
switch (cur_cpu_spec->pmc_type) {
|
|
#ifdef HAS_PPC_PMC_IBM
|
|
case PPC_PMC_IBM:
|
|
attrs = ibm_common_attrs;
|
|
nattrs = sizeof(ibm_common_attrs) / sizeof(struct device_attribute);
|
|
pmc_attrs = classic_pmc_attrs;
|
|
break;
|
|
#endif /* HAS_PPC_PMC_IBM */
|
|
#ifdef HAS_PPC_PMC_G4
|
|
case PPC_PMC_G4:
|
|
attrs = g4_common_attrs;
|
|
nattrs = sizeof(g4_common_attrs) / sizeof(struct device_attribute);
|
|
pmc_attrs = classic_pmc_attrs;
|
|
break;
|
|
#endif /* HAS_PPC_PMC_G4 */
|
|
#ifdef HAS_PPC_PMC_PA6T
|
|
case PPC_PMC_PA6T:
|
|
/* PA Semi starts counting at PMC0 */
|
|
attrs = pa6t_attrs;
|
|
nattrs = sizeof(pa6t_attrs) / sizeof(struct device_attribute);
|
|
pmc_attrs = NULL;
|
|
break;
|
|
#endif /* HAS_PPC_PMC_PA6T */
|
|
default:
|
|
attrs = NULL;
|
|
nattrs = 0;
|
|
pmc_attrs = NULL;
|
|
}
|
|
|
|
for (i = 0; i < nattrs; i++)
|
|
device_remove_file(s, &attrs[i]);
|
|
|
|
if (pmc_attrs)
|
|
for (i = 0; i < cur_cpu_spec->num_pmcs; i++)
|
|
device_remove_file(s, &pmc_attrs[i]);
|
|
|
|
#ifdef CONFIG_PPC64
|
|
if (cpu_has_feature(CPU_FTR_MMCRA))
|
|
device_remove_file(s, &dev_attr_mmcra);
|
|
|
|
if (cpu_has_feature(CPU_FTR_PURR))
|
|
device_remove_file(s, &dev_attr_purr);
|
|
|
|
if (cpu_has_feature(CPU_FTR_SPURR))
|
|
device_remove_file(s, &dev_attr_spurr);
|
|
|
|
if (cpu_has_feature(CPU_FTR_DSCR))
|
|
device_remove_file(s, &dev_attr_dscr);
|
|
|
|
if (cpu_has_feature(CPU_FTR_PPCAS_ARCH_V2))
|
|
device_remove_file(s, &dev_attr_pir);
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#ifdef CONFIG_PPC_FSL_BOOK3E
|
|
if (PVR_VER(cur_cpu_spec->pvr_value) == PVR_VER_E6500) {
|
|
device_remove_file(s, &dev_attr_pw20_state);
|
|
device_remove_file(s, &dev_attr_pw20_wait_time);
|
|
|
|
device_remove_file(s, &dev_attr_altivec_idle);
|
|
device_remove_file(s, &dev_attr_altivec_idle_wait_time);
|
|
}
|
|
#endif
|
|
cacheinfo_cpu_offline(cpu);
|
|
}
|
|
|
|
#ifdef CONFIG_ARCH_CPU_PROBE_RELEASE
|
|
ssize_t arch_cpu_probe(const char *buf, size_t count)
|
|
{
|
|
if (ppc_md.cpu_probe)
|
|
return ppc_md.cpu_probe(buf, count);
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
ssize_t arch_cpu_release(const char *buf, size_t count)
|
|
{
|
|
if (ppc_md.cpu_release)
|
|
return ppc_md.cpu_release(buf, count);
|
|
|
|
return -EINVAL;
|
|
}
|
|
#endif /* CONFIG_ARCH_CPU_PROBE_RELEASE */
|
|
|
|
#endif /* CONFIG_HOTPLUG_CPU */
|
|
|
|
static int sysfs_cpu_notify(struct notifier_block *self,
|
|
unsigned long action, void *hcpu)
|
|
{
|
|
unsigned int cpu = (unsigned int)(long)hcpu;
|
|
|
|
switch (action) {
|
|
case CPU_ONLINE:
|
|
case CPU_ONLINE_FROZEN:
|
|
register_cpu_online(cpu);
|
|
break;
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
case CPU_DEAD:
|
|
case CPU_DEAD_FROZEN:
|
|
unregister_cpu_online(cpu);
|
|
break;
|
|
#endif
|
|
}
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
static struct notifier_block sysfs_cpu_nb = {
|
|
.notifier_call = sysfs_cpu_notify,
|
|
};
|
|
|
|
static DEFINE_MUTEX(cpu_mutex);
|
|
|
|
int cpu_add_dev_attr(struct device_attribute *attr)
|
|
{
|
|
int cpu;
|
|
|
|
mutex_lock(&cpu_mutex);
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
device_create_file(get_cpu_device(cpu), attr);
|
|
}
|
|
|
|
mutex_unlock(&cpu_mutex);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(cpu_add_dev_attr);
|
|
|
|
int cpu_add_dev_attr_group(struct attribute_group *attrs)
|
|
{
|
|
int cpu;
|
|
struct device *dev;
|
|
int ret;
|
|
|
|
mutex_lock(&cpu_mutex);
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
dev = get_cpu_device(cpu);
|
|
ret = sysfs_create_group(&dev->kobj, attrs);
|
|
WARN_ON(ret != 0);
|
|
}
|
|
|
|
mutex_unlock(&cpu_mutex);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(cpu_add_dev_attr_group);
|
|
|
|
|
|
void cpu_remove_dev_attr(struct device_attribute *attr)
|
|
{
|
|
int cpu;
|
|
|
|
mutex_lock(&cpu_mutex);
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
device_remove_file(get_cpu_device(cpu), attr);
|
|
}
|
|
|
|
mutex_unlock(&cpu_mutex);
|
|
}
|
|
EXPORT_SYMBOL_GPL(cpu_remove_dev_attr);
|
|
|
|
void cpu_remove_dev_attr_group(struct attribute_group *attrs)
|
|
{
|
|
int cpu;
|
|
struct device *dev;
|
|
|
|
mutex_lock(&cpu_mutex);
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
dev = get_cpu_device(cpu);
|
|
sysfs_remove_group(&dev->kobj, attrs);
|
|
}
|
|
|
|
mutex_unlock(&cpu_mutex);
|
|
}
|
|
EXPORT_SYMBOL_GPL(cpu_remove_dev_attr_group);
|
|
|
|
|
|
/* NUMA stuff */
|
|
|
|
#ifdef CONFIG_NUMA
|
|
static void register_nodes(void)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_NUMNODES; i++)
|
|
register_one_node(i);
|
|
}
|
|
|
|
int sysfs_add_device_to_node(struct device *dev, int nid)
|
|
{
|
|
struct node *node = node_devices[nid];
|
|
return sysfs_create_link(&node->dev.kobj, &dev->kobj,
|
|
kobject_name(&dev->kobj));
|
|
}
|
|
EXPORT_SYMBOL_GPL(sysfs_add_device_to_node);
|
|
|
|
void sysfs_remove_device_from_node(struct device *dev, int nid)
|
|
{
|
|
struct node *node = node_devices[nid];
|
|
sysfs_remove_link(&node->dev.kobj, kobject_name(&dev->kobj));
|
|
}
|
|
EXPORT_SYMBOL_GPL(sysfs_remove_device_from_node);
|
|
|
|
#else
|
|
static void register_nodes(void)
|
|
{
|
|
return;
|
|
}
|
|
|
|
#endif
|
|
|
|
/* Only valid if CPU is present. */
|
|
static ssize_t show_physical_id(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct cpu *cpu = container_of(dev, struct cpu, dev);
|
|
|
|
return sprintf(buf, "%d\n", get_hard_smp_processor_id(cpu->dev.id));
|
|
}
|
|
static DEVICE_ATTR(physical_id, 0444, show_physical_id, NULL);
|
|
|
|
static int __init topology_init(void)
|
|
{
|
|
int cpu;
|
|
|
|
register_nodes();
|
|
|
|
cpu_notifier_register_begin();
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
struct cpu *c = &per_cpu(cpu_devices, cpu);
|
|
|
|
/*
|
|
* For now, we just see if the system supports making
|
|
* the RTAS calls for CPU hotplug. But, there may be a
|
|
* more comprehensive way to do this for an individual
|
|
* CPU. For instance, the boot cpu might never be valid
|
|
* for hotplugging.
|
|
*/
|
|
if (ppc_md.cpu_die)
|
|
c->hotpluggable = 1;
|
|
|
|
if (cpu_online(cpu) || c->hotpluggable) {
|
|
register_cpu(c, cpu);
|
|
|
|
device_create_file(&c->dev, &dev_attr_physical_id);
|
|
}
|
|
|
|
if (cpu_online(cpu))
|
|
register_cpu_online(cpu);
|
|
}
|
|
|
|
__register_cpu_notifier(&sysfs_cpu_nb);
|
|
|
|
cpu_notifier_register_done();
|
|
|
|
#ifdef CONFIG_PPC64
|
|
sysfs_create_dscr_default();
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
return 0;
|
|
}
|
|
subsys_initcall(topology_init);
|