c25620d766
Add platform specific SPU run control routines to the spufs. The current spufs implementation uses the SPU master run control bit (MFC_SR1[S]) to control SPE execution, but the PS3 hypervisor does not support the use of this feature. This change adds the run control wrapper routies spu_enable_spu() and spu_disable_spu(). The bare metal routines use the master run control bit, and the PS3 specific routines use the priv2 run control register. An outstanding enhancement for the PS3 would be to add a guard to check for incorrect access to the spu problem state when the spu context is disabled. This check could be implemented with a flag added to the spu context that would inhibit mapping problem state pages, and a routine to unmap spu problem state pages. When the spu is enabled with ps3_enable_spu() the flag would be set allowing pages to be mapped, and when the spu is disabled with ps3_disable_spu() the flag would be cleared and mapped problem state pages would be unmapped. Signed-off-by: Masato Noguchi <Masato.Noguchi@jp.sony.com> Signed-off-by: Geoff Levand <geoffrey.levand@am.sony.com> Signed-off-by: Jeremy Kerr <jk@ozlabs.org> Signed-off-by: Paul Mackerras <paulus@samba.org>
557 lines
13 KiB
C
557 lines
13 KiB
C
/*
|
|
* spu management operations for of based platforms
|
|
*
|
|
* (C) Copyright IBM Deutschland Entwicklung GmbH 2005
|
|
* Copyright 2006 Sony Corp.
|
|
* (C) Copyright 2007 TOSHIBA CORPORATION
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; version 2 of the License.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with this program; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/list.h>
|
|
#include <linux/module.h>
|
|
#include <linux/ptrace.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/io.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/device.h>
|
|
|
|
#include <asm/spu.h>
|
|
#include <asm/spu_priv1.h>
|
|
#include <asm/firmware.h>
|
|
#include <asm/prom.h>
|
|
|
|
#include "spufs/spufs.h"
|
|
#include "interrupt.h"
|
|
|
|
struct device_node *spu_devnode(struct spu *spu)
|
|
{
|
|
return spu->devnode;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(spu_devnode);
|
|
|
|
static u64 __init find_spu_unit_number(struct device_node *spe)
|
|
{
|
|
const unsigned int *prop;
|
|
int proplen;
|
|
|
|
/* new device trees should provide the physical-id attribute */
|
|
prop = of_get_property(spe, "physical-id", &proplen);
|
|
if (proplen == 4)
|
|
return (u64)*prop;
|
|
|
|
/* celleb device tree provides the unit-id */
|
|
prop = of_get_property(spe, "unit-id", &proplen);
|
|
if (proplen == 4)
|
|
return (u64)*prop;
|
|
|
|
/* legacy device trees provide the id in the reg attribute */
|
|
prop = of_get_property(spe, "reg", &proplen);
|
|
if (proplen == 4)
|
|
return (u64)*prop;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void spu_unmap(struct spu *spu)
|
|
{
|
|
if (!firmware_has_feature(FW_FEATURE_LPAR))
|
|
iounmap(spu->priv1);
|
|
iounmap(spu->priv2);
|
|
iounmap(spu->problem);
|
|
iounmap((__force u8 __iomem *)spu->local_store);
|
|
}
|
|
|
|
static int __init spu_map_interrupts_old(struct spu *spu,
|
|
struct device_node *np)
|
|
{
|
|
unsigned int isrc;
|
|
const u32 *tmp;
|
|
int nid;
|
|
|
|
/* Get the interrupt source unit from the device-tree */
|
|
tmp = of_get_property(np, "isrc", NULL);
|
|
if (!tmp)
|
|
return -ENODEV;
|
|
isrc = tmp[0];
|
|
|
|
tmp = of_get_property(np->parent->parent, "node-id", NULL);
|
|
if (!tmp) {
|
|
printk(KERN_WARNING "%s: can't find node-id\n", __FUNCTION__);
|
|
nid = spu->node;
|
|
} else
|
|
nid = tmp[0];
|
|
|
|
/* Add the node number */
|
|
isrc |= nid << IIC_IRQ_NODE_SHIFT;
|
|
|
|
/* Now map interrupts of all 3 classes */
|
|
spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
|
|
spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
|
|
spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);
|
|
|
|
/* Right now, we only fail if class 2 failed */
|
|
return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
|
|
}
|
|
|
|
static void __iomem * __init spu_map_prop_old(struct spu *spu,
|
|
struct device_node *n,
|
|
const char *name)
|
|
{
|
|
const struct address_prop {
|
|
unsigned long address;
|
|
unsigned int len;
|
|
} __attribute__((packed)) *prop;
|
|
int proplen;
|
|
|
|
prop = of_get_property(n, name, &proplen);
|
|
if (prop == NULL || proplen != sizeof (struct address_prop))
|
|
return NULL;
|
|
|
|
return ioremap(prop->address, prop->len);
|
|
}
|
|
|
|
static int __init spu_map_device_old(struct spu *spu)
|
|
{
|
|
struct device_node *node = spu->devnode;
|
|
const char *prop;
|
|
int ret;
|
|
|
|
ret = -ENODEV;
|
|
spu->name = of_get_property(node, "name", NULL);
|
|
if (!spu->name)
|
|
goto out;
|
|
|
|
prop = of_get_property(node, "local-store", NULL);
|
|
if (!prop)
|
|
goto out;
|
|
spu->local_store_phys = *(unsigned long *)prop;
|
|
|
|
/* we use local store as ram, not io memory */
|
|
spu->local_store = (void __force *)
|
|
spu_map_prop_old(spu, node, "local-store");
|
|
if (!spu->local_store)
|
|
goto out;
|
|
|
|
prop = of_get_property(node, "problem", NULL);
|
|
if (!prop)
|
|
goto out_unmap;
|
|
spu->problem_phys = *(unsigned long *)prop;
|
|
|
|
spu->problem = spu_map_prop_old(spu, node, "problem");
|
|
if (!spu->problem)
|
|
goto out_unmap;
|
|
|
|
spu->priv2 = spu_map_prop_old(spu, node, "priv2");
|
|
if (!spu->priv2)
|
|
goto out_unmap;
|
|
|
|
if (!firmware_has_feature(FW_FEATURE_LPAR)) {
|
|
spu->priv1 = spu_map_prop_old(spu, node, "priv1");
|
|
if (!spu->priv1)
|
|
goto out_unmap;
|
|
}
|
|
|
|
ret = 0;
|
|
goto out;
|
|
|
|
out_unmap:
|
|
spu_unmap(spu);
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
|
|
{
|
|
struct of_irq oirq;
|
|
int ret;
|
|
int i;
|
|
|
|
for (i=0; i < 3; i++) {
|
|
ret = of_irq_map_one(np, i, &oirq);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to get irq %d\n", i);
|
|
goto err;
|
|
}
|
|
ret = -EINVAL;
|
|
pr_debug(" irq %d no 0x%x on %s\n", i, oirq.specifier[0],
|
|
oirq.controller->full_name);
|
|
spu->irqs[i] = irq_create_of_mapping(oirq.controller,
|
|
oirq.specifier, oirq.size);
|
|
if (spu->irqs[i] == NO_IRQ) {
|
|
pr_debug("spu_new: failed to map it !\n");
|
|
goto err;
|
|
}
|
|
}
|
|
return 0;
|
|
|
|
err:
|
|
pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier,
|
|
spu->name);
|
|
for (; i >= 0; i--) {
|
|
if (spu->irqs[i] != NO_IRQ)
|
|
irq_dispose_mapping(spu->irqs[i]);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int spu_map_resource(struct spu *spu, int nr,
|
|
void __iomem** virt, unsigned long *phys)
|
|
{
|
|
struct device_node *np = spu->devnode;
|
|
struct resource resource = { };
|
|
unsigned long len;
|
|
int ret;
|
|
|
|
ret = of_address_to_resource(np, nr, &resource);
|
|
if (ret)
|
|
return ret;
|
|
if (phys)
|
|
*phys = resource.start;
|
|
len = resource.end - resource.start + 1;
|
|
*virt = ioremap(resource.start, len);
|
|
if (!*virt)
|
|
return -EINVAL;
|
|
return 0;
|
|
}
|
|
|
|
static int __init spu_map_device(struct spu *spu)
|
|
{
|
|
struct device_node *np = spu->devnode;
|
|
int ret = -ENODEV;
|
|
|
|
spu->name = of_get_property(np, "name", NULL);
|
|
if (!spu->name)
|
|
goto out;
|
|
|
|
ret = spu_map_resource(spu, 0, (void __iomem**)&spu->local_store,
|
|
&spu->local_store_phys);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 0\n",
|
|
np->full_name);
|
|
goto out;
|
|
}
|
|
ret = spu_map_resource(spu, 1, (void __iomem**)&spu->problem,
|
|
&spu->problem_phys);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 1\n",
|
|
np->full_name);
|
|
goto out_unmap;
|
|
}
|
|
ret = spu_map_resource(spu, 2, (void __iomem**)&spu->priv2, NULL);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 2\n",
|
|
np->full_name);
|
|
goto out_unmap;
|
|
}
|
|
if (!firmware_has_feature(FW_FEATURE_LPAR))
|
|
ret = spu_map_resource(spu, 3,
|
|
(void __iomem**)&spu->priv1, NULL);
|
|
if (ret) {
|
|
pr_debug("spu_new: failed to map %s resource 3\n",
|
|
np->full_name);
|
|
goto out_unmap;
|
|
}
|
|
pr_debug("spu_new: %s maps:\n", np->full_name);
|
|
pr_debug(" local store : 0x%016lx -> 0x%p\n",
|
|
spu->local_store_phys, spu->local_store);
|
|
pr_debug(" problem state : 0x%016lx -> 0x%p\n",
|
|
spu->problem_phys, spu->problem);
|
|
pr_debug(" priv2 : 0x%p\n", spu->priv2);
|
|
pr_debug(" priv1 : 0x%p\n", spu->priv1);
|
|
|
|
return 0;
|
|
|
|
out_unmap:
|
|
spu_unmap(spu);
|
|
out:
|
|
pr_debug("failed to map spe %s: %d\n", spu->name, ret);
|
|
return ret;
|
|
}
|
|
|
|
static int __init of_enumerate_spus(int (*fn)(void *data))
|
|
{
|
|
int ret;
|
|
struct device_node *node;
|
|
unsigned int n = 0;
|
|
|
|
ret = -ENODEV;
|
|
for (node = of_find_node_by_type(NULL, "spe");
|
|
node; node = of_find_node_by_type(node, "spe")) {
|
|
ret = fn(node);
|
|
if (ret) {
|
|
printk(KERN_WARNING "%s: Error initializing %s\n",
|
|
__FUNCTION__, node->name);
|
|
break;
|
|
}
|
|
n++;
|
|
}
|
|
return ret ? ret : n;
|
|
}
|
|
|
|
static int __init of_create_spu(struct spu *spu, void *data)
|
|
{
|
|
int ret;
|
|
struct device_node *spe = (struct device_node *)data;
|
|
static int legacy_map = 0, legacy_irq = 0;
|
|
|
|
spu->devnode = of_node_get(spe);
|
|
spu->spe_id = find_spu_unit_number(spe);
|
|
|
|
spu->node = of_node_to_nid(spe);
|
|
if (spu->node >= MAX_NUMNODES) {
|
|
printk(KERN_WARNING "SPE %s on node %d ignored,"
|
|
" node number too big\n", spe->full_name, spu->node);
|
|
printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
|
|
ret = -ENODEV;
|
|
goto out;
|
|
}
|
|
|
|
ret = spu_map_device(spu);
|
|
if (ret) {
|
|
if (!legacy_map) {
|
|
legacy_map = 1;
|
|
printk(KERN_WARNING "%s: Legacy device tree found, "
|
|
"trying to map old style\n", __FUNCTION__);
|
|
}
|
|
ret = spu_map_device_old(spu);
|
|
if (ret) {
|
|
printk(KERN_ERR "Unable to map %s\n",
|
|
spu->name);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
ret = spu_map_interrupts(spu, spe);
|
|
if (ret) {
|
|
if (!legacy_irq) {
|
|
legacy_irq = 1;
|
|
printk(KERN_WARNING "%s: Legacy device tree found, "
|
|
"trying old style irq\n", __FUNCTION__);
|
|
}
|
|
ret = spu_map_interrupts_old(spu, spe);
|
|
if (ret) {
|
|
printk(KERN_ERR "%s: could not map interrupts\n",
|
|
spu->name);
|
|
goto out_unmap;
|
|
}
|
|
}
|
|
|
|
pr_debug("Using SPE %s %p %p %p %p %d\n", spu->name,
|
|
spu->local_store, spu->problem, spu->priv1,
|
|
spu->priv2, spu->number);
|
|
goto out;
|
|
|
|
out_unmap:
|
|
spu_unmap(spu);
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static int of_destroy_spu(struct spu *spu)
|
|
{
|
|
spu_unmap(spu);
|
|
of_node_put(spu->devnode);
|
|
return 0;
|
|
}
|
|
|
|
static void enable_spu_by_master_run(struct spu_context *ctx)
|
|
{
|
|
ctx->ops->master_start(ctx);
|
|
}
|
|
|
|
static void disable_spu_by_master_run(struct spu_context *ctx)
|
|
{
|
|
ctx->ops->master_stop(ctx);
|
|
}
|
|
|
|
/* Hardcoded affinity idxs for qs20 */
|
|
#define QS20_SPES_PER_BE 8
|
|
static int qs20_reg_idxs[QS20_SPES_PER_BE] = { 0, 2, 4, 6, 7, 5, 3, 1 };
|
|
static int qs20_reg_memory[QS20_SPES_PER_BE] = { 1, 1, 0, 0, 0, 0, 0, 0 };
|
|
|
|
static struct spu *spu_lookup_reg(int node, u32 reg)
|
|
{
|
|
struct spu *spu;
|
|
const u32 *spu_reg;
|
|
|
|
list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
|
|
spu_reg = of_get_property(spu_devnode(spu), "reg", NULL);
|
|
if (*spu_reg == reg)
|
|
return spu;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static void init_affinity_qs20_harcoded(void)
|
|
{
|
|
int node, i;
|
|
struct spu *last_spu, *spu;
|
|
u32 reg;
|
|
|
|
for (node = 0; node < MAX_NUMNODES; node++) {
|
|
last_spu = NULL;
|
|
for (i = 0; i < QS20_SPES_PER_BE; i++) {
|
|
reg = qs20_reg_idxs[i];
|
|
spu = spu_lookup_reg(node, reg);
|
|
if (!spu)
|
|
continue;
|
|
spu->has_mem_affinity = qs20_reg_memory[reg];
|
|
if (last_spu)
|
|
list_add_tail(&spu->aff_list,
|
|
&last_spu->aff_list);
|
|
last_spu = spu;
|
|
}
|
|
}
|
|
}
|
|
|
|
static int of_has_vicinity(void)
|
|
{
|
|
struct device_node *dn;
|
|
|
|
for_each_node_by_type(dn, "spe") {
|
|
if (of_find_property(dn, "vicinity", NULL)) {
|
|
of_node_put(dn);
|
|
return 1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static struct spu *devnode_spu(int cbe, struct device_node *dn)
|
|
{
|
|
struct spu *spu;
|
|
|
|
list_for_each_entry(spu, &cbe_spu_info[cbe].spus, cbe_list)
|
|
if (spu_devnode(spu) == dn)
|
|
return spu;
|
|
return NULL;
|
|
}
|
|
|
|
static struct spu *
|
|
neighbour_spu(int cbe, struct device_node *target, struct device_node *avoid)
|
|
{
|
|
struct spu *spu;
|
|
struct device_node *spu_dn;
|
|
const phandle *vic_handles;
|
|
int lenp, i;
|
|
|
|
list_for_each_entry(spu, &cbe_spu_info[cbe].spus, cbe_list) {
|
|
spu_dn = spu_devnode(spu);
|
|
if (spu_dn == avoid)
|
|
continue;
|
|
vic_handles = of_get_property(spu_dn, "vicinity", &lenp);
|
|
for (i=0; i < (lenp / sizeof(phandle)); i++) {
|
|
if (vic_handles[i] == target->linux_phandle)
|
|
return spu;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static void init_affinity_node(int cbe)
|
|
{
|
|
struct spu *spu, *last_spu;
|
|
struct device_node *vic_dn, *last_spu_dn;
|
|
phandle avoid_ph;
|
|
const phandle *vic_handles;
|
|
const char *name;
|
|
int lenp, i, added;
|
|
|
|
last_spu = list_first_entry(&cbe_spu_info[cbe].spus, struct spu,
|
|
cbe_list);
|
|
avoid_ph = 0;
|
|
for (added = 1; added < cbe_spu_info[cbe].n_spus; added++) {
|
|
last_spu_dn = spu_devnode(last_spu);
|
|
vic_handles = of_get_property(last_spu_dn, "vicinity", &lenp);
|
|
|
|
/*
|
|
* Walk through each phandle in vicinity property of the spu
|
|
* (tipically two vicinity phandles per spe node)
|
|
*/
|
|
for (i = 0; i < (lenp / sizeof(phandle)); i++) {
|
|
if (vic_handles[i] == avoid_ph)
|
|
continue;
|
|
|
|
vic_dn = of_find_node_by_phandle(vic_handles[i]);
|
|
if (!vic_dn)
|
|
continue;
|
|
|
|
/* a neighbour might be spe, mic-tm, or bif0 */
|
|
name = of_get_property(vic_dn, "name", NULL);
|
|
if (!name)
|
|
continue;
|
|
|
|
if (strcmp(name, "spe") == 0) {
|
|
spu = devnode_spu(cbe, vic_dn);
|
|
avoid_ph = last_spu_dn->linux_phandle;
|
|
} else {
|
|
/*
|
|
* "mic-tm" and "bif0" nodes do not have
|
|
* vicinity property. So we need to find the
|
|
* spe which has vic_dn as neighbour, but
|
|
* skipping the one we came from (last_spu_dn)
|
|
*/
|
|
spu = neighbour_spu(cbe, vic_dn, last_spu_dn);
|
|
if (!spu)
|
|
continue;
|
|
if (!strcmp(name, "mic-tm")) {
|
|
last_spu->has_mem_affinity = 1;
|
|
spu->has_mem_affinity = 1;
|
|
}
|
|
avoid_ph = vic_dn->linux_phandle;
|
|
}
|
|
|
|
list_add_tail(&spu->aff_list, &last_spu->aff_list);
|
|
last_spu = spu;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void init_affinity_fw(void)
|
|
{
|
|
int cbe;
|
|
|
|
for (cbe = 0; cbe < MAX_NUMNODES; cbe++)
|
|
init_affinity_node(cbe);
|
|
}
|
|
|
|
static int __init init_affinity(void)
|
|
{
|
|
if (of_has_vicinity()) {
|
|
init_affinity_fw();
|
|
} else {
|
|
long root = of_get_flat_dt_root();
|
|
if (of_flat_dt_is_compatible(root, "IBM,CPBW-1.0"))
|
|
init_affinity_qs20_harcoded();
|
|
else
|
|
printk("No affinity configuration found\n");
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
const struct spu_management_ops spu_management_of_ops = {
|
|
.enumerate_spus = of_enumerate_spus,
|
|
.create_spu = of_create_spu,
|
|
.destroy_spu = of_destroy_spu,
|
|
.enable_spu = enable_spu_by_master_run,
|
|
.disable_spu = disable_spu_by_master_run,
|
|
.init_affinity = init_affinity,
|
|
};
|