forked from Minki/linux
ccb4911598
The difference between 'nid' and 'node' fields in an spu structure was used incorrectly. The common 'node' number now reflects the NUMA node, and it is used in other places in the code as well. The 'nid' value is meaningful only in one place, namely the computation of the interrupt numbers based on the physical location of an spu. Consequently, we look it up directly in the place where it is used now. Signed-off-by: Arnd Bergmann <arnd.bergmann@de.ibm.com>
549 lines
13 KiB
C
549 lines
13 KiB
C
/*
|
|
* spu hypervisor abstraction for direct hardware access.
|
|
*
|
|
* (C) Copyright IBM Deutschland Entwicklung GmbH 2005
|
|
* Copyright 2006 Sony Corp.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; version 2 of the License.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/list.h>
|
|
#include <linux/module.h>
|
|
#include <linux/ptrace.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/io.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/device.h>
|
|
|
|
#include <asm/spu.h>
|
|
#include <asm/spu_priv1.h>
|
|
#include <asm/firmware.h>
|
|
#include <asm/prom.h>
|
|
|
|
#include "interrupt.h"
|
|
#include "spu_priv1_mmio.h"
|
|
|
|
static DEFINE_MUTEX(add_spumem_mutex);
|
|
|
|
struct spu_pdata {
|
|
struct device_node *devnode;
|
|
struct spu_priv1 __iomem *priv1;
|
|
};
|
|
|
|
static struct spu_pdata *spu_get_pdata(struct spu *spu)
|
|
{
|
|
BUG_ON(!spu->pdata);
|
|
return spu->pdata;
|
|
}
|
|
|
|
struct device_node *spu_devnode(struct spu *spu)
|
|
{
|
|
return spu_get_pdata(spu)->devnode;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(spu_devnode);
|
|
|
|
static int __init cell_spuprop_present(struct spu *spu, struct device_node *spe,
|
|
const char *prop)
|
|
{
|
|
const struct address_prop {
|
|
unsigned long address;
|
|
unsigned int len;
|
|
} __attribute__((packed)) *p;
|
|
int proplen;
|
|
|
|
unsigned long start_pfn, nr_pages;
|
|
struct pglist_data *pgdata;
|
|
struct zone *zone;
|
|
int ret;
|
|
|
|
p = get_property(spe, prop, &proplen);
|
|
WARN_ON(proplen != sizeof (*p));
|
|
|
|
start_pfn = p->address >> PAGE_SHIFT;
|
|
nr_pages = ((unsigned long)p->len + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
|
|
pgdata = NODE_DATA(spu->node);
|
|
zone = pgdata->node_zones;
|
|
|
|
/* XXX rethink locking here */
|
|
mutex_lock(&add_spumem_mutex);
|
|
ret = __add_pages(zone, start_pfn, nr_pages);
|
|
mutex_unlock(&add_spumem_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void __iomem * __init map_spe_prop(struct spu *spu,
|
|
struct device_node *n, const char *name)
|
|
{
|
|
const struct address_prop {
|
|
unsigned long address;
|
|
unsigned int len;
|
|
} __attribute__((packed)) *prop;
|
|
|
|
const void *p;
|
|
int proplen;
|
|
void __iomem *ret = NULL;
|
|
int err = 0;
|
|
|
|
p = get_property(n, name, &proplen);
|
|
if (proplen != sizeof (struct address_prop))
|
|
return NULL;
|
|
|
|
prop = p;
|
|
|
|
err = cell_spuprop_present(spu, n, name);
|
|
if (err && (err != -EEXIST))
|
|
goto out;
|
|
|
|
ret = ioremap(prop->address, prop->len);
|
|
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static void spu_unmap(struct spu *spu)
|
|
{
|
|
iounmap(spu->priv2);
|
|
iounmap(spu_get_pdata(spu)->priv1);
|
|
iounmap(spu->problem);
|
|
iounmap((__force u8 __iomem *)spu->local_store);
|
|
}
|
|
|
|
static int __init spu_map_interrupts_old(struct spu *spu,
|
|
struct device_node *np)
|
|
{
|
|
unsigned int isrc;
|
|
const u32 *tmp;
|
|
int nid;
|
|
|
|
/* Get the interrupt source unit from the device-tree */
|
|
tmp = get_property(np, "isrc", NULL);
|
|
if (!tmp)
|
|
return -ENODEV;
|
|
isrc = tmp[0];
|
|
|
|
tmp = get_property(np->parent->parent, "node-id", NULL);
|
|
if (!tmp) {
|
|
printk(KERN_WARNING "%s: can't find node-id\n", __FUNCTION__);
|
|
nid = spu->node;
|
|
} else
|
|
nid = tmp[0];
|
|
|
|
/* Add the node number */
|
|
isrc |= nid << IIC_IRQ_NODE_SHIFT;
|
|
|
|
/* Now map interrupts of all 3 classes */
|
|
spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
|
|
spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
|
|
spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);
|
|
|
|
/* Right now, we only fail if class 2 failed */
|
|
return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
|
|
}
|
|
|
|
static int __init spu_map_device_old(struct spu *spu, struct device_node *node)
|
|
{
|
|
const char *prop;
|
|
int ret;
|
|
|
|
ret = -ENODEV;
|
|
spu->name = get_property(node, "name", NULL);
|
|
if (!spu->name)
|
|
goto out;
|
|
|
|
prop = get_property(node, "local-store", NULL);
|
|
if (!prop)
|
|
goto out;
|
|
spu->local_store_phys = *(unsigned long *)prop;
|
|
|
|
/* we use local store as ram, not io memory */
|
|
spu->local_store = (void __force *)
|
|
map_spe_prop(spu, node, "local-store");
|
|
if (!spu->local_store)
|
|
goto out;
|
|
|
|
prop = get_property(node, "problem", NULL);
|
|
if (!prop)
|
|
goto out_unmap;
|
|
spu->problem_phys = *(unsigned long *)prop;
|
|
|
|
spu->problem= map_spe_prop(spu, node, "problem");
|
|
if (!spu->problem)
|
|
goto out_unmap;
|
|
|
|
spu_get_pdata(spu)->priv1= map_spe_prop(spu, node, "priv1");
|
|
|
|
spu->priv2= map_spe_prop(spu, node, "priv2");
|
|
if (!spu->priv2)
|
|
goto out_unmap;
|
|
ret = 0;
|
|
goto out;
|
|
|
|
out_unmap:
|
|
spu_unmap(spu);
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
|
|
{
|
|
struct of_irq oirq;
|
|
int ret;
|
|
int i;
|
|
|
|
for (i=0; i < 3; i++) {
|
|
ret = of_irq_map_one(np, i, &oirq);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to get irq %d\n", i);
|
|
goto err;
|
|
}
|
|
ret = -EINVAL;
|
|
pr_debug(" irq %d no 0x%x on %s\n", i, oirq.specifier[0],
|
|
oirq.controller->full_name);
|
|
spu->irqs[i] = irq_create_of_mapping(oirq.controller,
|
|
oirq.specifier, oirq.size);
|
|
if (spu->irqs[i] == NO_IRQ) {
|
|
pr_debug("spu_new: failed to map it !\n");
|
|
goto err;
|
|
}
|
|
}
|
|
return 0;
|
|
|
|
err:
|
|
pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier,
|
|
spu->name);
|
|
for (; i >= 0; i--) {
|
|
if (spu->irqs[i] != NO_IRQ)
|
|
irq_dispose_mapping(spu->irqs[i]);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int spu_map_resource(struct spu *spu, int nr,
|
|
void __iomem** virt, unsigned long *phys)
|
|
{
|
|
struct device_node *np = spu_get_pdata(spu)->devnode;
|
|
unsigned long start_pfn, nr_pages;
|
|
struct pglist_data *pgdata;
|
|
struct zone *zone;
|
|
struct resource resource = { };
|
|
unsigned long len;
|
|
int ret;
|
|
|
|
ret = of_address_to_resource(np, nr, &resource);
|
|
if (ret)
|
|
goto out;
|
|
|
|
if (phys)
|
|
*phys = resource.start;
|
|
len = resource.end - resource.start + 1;
|
|
*virt = ioremap(resource.start, len);
|
|
if (!*virt)
|
|
ret = -EINVAL;
|
|
|
|
start_pfn = resource.start >> PAGE_SHIFT;
|
|
nr_pages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
|
|
pgdata = NODE_DATA(spu->node);
|
|
zone = pgdata->node_zones;
|
|
|
|
/* XXX rethink locking here */
|
|
mutex_lock(&add_spumem_mutex);
|
|
ret = __add_pages(zone, start_pfn, nr_pages);
|
|
mutex_unlock(&add_spumem_mutex);
|
|
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static int __init spu_map_device(struct spu *spu)
|
|
{
|
|
struct device_node *np = spu_get_pdata(spu)->devnode;
|
|
int ret = -ENODEV;
|
|
|
|
spu->name = get_property(np, "name", NULL);
|
|
if (!spu->name)
|
|
goto out;
|
|
|
|
ret = spu_map_resource(spu, 0, (void __iomem**)&spu->local_store,
|
|
&spu->local_store_phys);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 0\n",
|
|
np->full_name);
|
|
goto out;
|
|
}
|
|
ret = spu_map_resource(spu, 1, (void __iomem**)&spu->problem,
|
|
&spu->problem_phys);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 1\n",
|
|
np->full_name);
|
|
goto out_unmap;
|
|
}
|
|
ret = spu_map_resource(spu, 2, (void __iomem**)&spu->priv2, NULL);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 2\n",
|
|
np->full_name);
|
|
goto out_unmap;
|
|
}
|
|
if (!firmware_has_feature(FW_FEATURE_LPAR))
|
|
ret = spu_map_resource(spu, 3,
|
|
(void __iomem**)&spu_get_pdata(spu)->priv1, NULL);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 3\n",
|
|
np->full_name);
|
|
goto out_unmap;
|
|
}
|
|
pr_debug("spu_new: %s maps:\n", np->full_name);
|
|
pr_debug(" local store : 0x%016lx -> 0x%p\n",
|
|
spu->local_store_phys, spu->local_store);
|
|
pr_debug(" problem state : 0x%016lx -> 0x%p\n",
|
|
spu->problem_phys, spu->problem);
|
|
pr_debug(" priv2 : 0x%p\n", spu->priv2);
|
|
pr_debug(" priv1 : 0x%p\n",
|
|
spu_get_pdata(spu)->priv1);
|
|
|
|
return 0;
|
|
|
|
out_unmap:
|
|
spu_unmap(spu);
|
|
out:
|
|
pr_debug("failed to map spe %s: %d\n", spu->name, ret);
|
|
return ret;
|
|
}
|
|
|
|
static int __init of_enumerate_spus(int (*fn)(void *data))
|
|
{
|
|
int ret;
|
|
struct device_node *node;
|
|
|
|
ret = -ENODEV;
|
|
for (node = of_find_node_by_type(NULL, "spe");
|
|
node; node = of_find_node_by_type(node, "spe")) {
|
|
ret = fn(node);
|
|
if (ret) {
|
|
printk(KERN_WARNING "%s: Error initializing %s\n",
|
|
__FUNCTION__, node->name);
|
|
break;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int __init of_create_spu(struct spu *spu, void *data)
|
|
{
|
|
int ret;
|
|
struct device_node *spe = (struct device_node *)data;
|
|
|
|
spu->pdata = kzalloc(sizeof(struct spu_pdata),
|
|
GFP_KERNEL);
|
|
if (!spu->pdata) {
|
|
ret = -ENOMEM;
|
|
goto out;
|
|
}
|
|
spu_get_pdata(spu)->devnode = of_node_get(spe);
|
|
|
|
spu->node = of_node_to_nid(spe);
|
|
if (spu->node >= MAX_NUMNODES) {
|
|
printk(KERN_WARNING "SPE %s on node %d ignored,"
|
|
" node number too big\n", spe->full_name, spu->node);
|
|
printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
|
|
ret = -ENODEV;
|
|
goto out_free;
|
|
}
|
|
|
|
ret = spu_map_device(spu);
|
|
/* try old method */
|
|
if (ret)
|
|
ret = spu_map_device_old(spu, spe);
|
|
if (ret)
|
|
goto out_free;
|
|
|
|
ret = spu_map_interrupts(spu, spe);
|
|
if (ret)
|
|
ret = spu_map_interrupts_old(spu, spe);
|
|
if (ret)
|
|
goto out_unmap;
|
|
|
|
pr_debug(KERN_DEBUG "Using SPE %s %p %p %p %p %d\n", spu->name,
|
|
spu->local_store, spu->problem, spu_get_pdata(spu)->priv1,
|
|
spu->priv2, spu->number);
|
|
goto out;
|
|
|
|
out_unmap:
|
|
spu_unmap(spu);
|
|
out_free:
|
|
kfree(spu->pdata);
|
|
spu->pdata = NULL;
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static int of_destroy_spu(struct spu *spu)
|
|
{
|
|
spu_unmap(spu);
|
|
of_node_put(spu_get_pdata(spu)->devnode);
|
|
kfree(spu->pdata);
|
|
spu->pdata = NULL;
|
|
return 0;
|
|
}
|
|
|
|
const struct spu_management_ops spu_management_of_ops = {
|
|
.enumerate_spus = of_enumerate_spus,
|
|
.create_spu = of_create_spu,
|
|
.destroy_spu = of_destroy_spu,
|
|
};
|
|
|
|
static void int_mask_and(struct spu *spu, int class, u64 mask)
|
|
{
|
|
u64 old_mask;
|
|
|
|
old_mask = in_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class]);
|
|
out_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class],
|
|
old_mask & mask);
|
|
}
|
|
|
|
static void int_mask_or(struct spu *spu, int class, u64 mask)
|
|
{
|
|
u64 old_mask;
|
|
|
|
old_mask = in_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class]);
|
|
out_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class],
|
|
old_mask | mask);
|
|
}
|
|
|
|
static void int_mask_set(struct spu *spu, int class, u64 mask)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class], mask);
|
|
}
|
|
|
|
static u64 int_mask_get(struct spu *spu, int class)
|
|
{
|
|
return in_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class]);
|
|
}
|
|
|
|
static void int_stat_clear(struct spu *spu, int class, u64 stat)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->int_stat_RW[class], stat);
|
|
}
|
|
|
|
static u64 int_stat_get(struct spu *spu, int class)
|
|
{
|
|
return in_be64(&spu_get_pdata(spu)->priv1->int_stat_RW[class]);
|
|
}
|
|
|
|
static void cpu_affinity_set(struct spu *spu, int cpu)
|
|
{
|
|
u64 target = iic_get_target_id(cpu);
|
|
u64 route = target << 48 | target << 32 | target << 16;
|
|
out_be64(&spu_get_pdata(spu)->priv1->int_route_RW, route);
|
|
}
|
|
|
|
static u64 mfc_dar_get(struct spu *spu)
|
|
{
|
|
return in_be64(&spu_get_pdata(spu)->priv1->mfc_dar_RW);
|
|
}
|
|
|
|
static u64 mfc_dsisr_get(struct spu *spu)
|
|
{
|
|
return in_be64(&spu_get_pdata(spu)->priv1->mfc_dsisr_RW);
|
|
}
|
|
|
|
static void mfc_dsisr_set(struct spu *spu, u64 dsisr)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->mfc_dsisr_RW, dsisr);
|
|
}
|
|
|
|
static void mfc_sdr_setup(struct spu *spu)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->mfc_sdr_RW, mfspr(SPRN_SDR1));
|
|
}
|
|
|
|
static void mfc_sr1_set(struct spu *spu, u64 sr1)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->mfc_sr1_RW, sr1);
|
|
}
|
|
|
|
static u64 mfc_sr1_get(struct spu *spu)
|
|
{
|
|
return in_be64(&spu_get_pdata(spu)->priv1->mfc_sr1_RW);
|
|
}
|
|
|
|
static void mfc_tclass_id_set(struct spu *spu, u64 tclass_id)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->mfc_tclass_id_RW, tclass_id);
|
|
}
|
|
|
|
static u64 mfc_tclass_id_get(struct spu *spu)
|
|
{
|
|
return in_be64(&spu_get_pdata(spu)->priv1->mfc_tclass_id_RW);
|
|
}
|
|
|
|
static void tlb_invalidate(struct spu *spu)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->tlb_invalidate_entry_W, 0ul);
|
|
}
|
|
|
|
static void resource_allocation_groupID_set(struct spu *spu, u64 id)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->resource_allocation_groupID_RW,
|
|
id);
|
|
}
|
|
|
|
static u64 resource_allocation_groupID_get(struct spu *spu)
|
|
{
|
|
return in_be64(
|
|
&spu_get_pdata(spu)->priv1->resource_allocation_groupID_RW);
|
|
}
|
|
|
|
static void resource_allocation_enable_set(struct spu *spu, u64 enable)
|
|
{
|
|
out_be64(&spu_get_pdata(spu)->priv1->resource_allocation_enable_RW,
|
|
enable);
|
|
}
|
|
|
|
static u64 resource_allocation_enable_get(struct spu *spu)
|
|
{
|
|
return in_be64(
|
|
&spu_get_pdata(spu)->priv1->resource_allocation_enable_RW);
|
|
}
|
|
|
|
const struct spu_priv1_ops spu_priv1_mmio_ops =
|
|
{
|
|
.int_mask_and = int_mask_and,
|
|
.int_mask_or = int_mask_or,
|
|
.int_mask_set = int_mask_set,
|
|
.int_mask_get = int_mask_get,
|
|
.int_stat_clear = int_stat_clear,
|
|
.int_stat_get = int_stat_get,
|
|
.cpu_affinity_set = cpu_affinity_set,
|
|
.mfc_dar_get = mfc_dar_get,
|
|
.mfc_dsisr_get = mfc_dsisr_get,
|
|
.mfc_dsisr_set = mfc_dsisr_set,
|
|
.mfc_sdr_setup = mfc_sdr_setup,
|
|
.mfc_sr1_set = mfc_sr1_set,
|
|
.mfc_sr1_get = mfc_sr1_get,
|
|
.mfc_tclass_id_set = mfc_tclass_id_set,
|
|
.mfc_tclass_id_get = mfc_tclass_id_get,
|
|
.tlb_invalidate = tlb_invalidate,
|
|
.resource_allocation_groupID_set = resource_allocation_groupID_set,
|
|
.resource_allocation_groupID_get = resource_allocation_groupID_get,
|
|
.resource_allocation_enable_set = resource_allocation_enable_set,
|
|
.resource_allocation_enable_get = resource_allocation_enable_get,
|
|
};
|