[Cbe-oss-dev] [PATCH] cell: abstract spu management routines
Benjamin Herrenschmidt
benh at kernel.crashing.org
Wed Nov 8 16:31:35 EST 2006
> For the IBM Cell Blade support, I put the hypervisor only resources that were
> in struct spu into a platform specific data structure struct platform_data.
We have no hypervisor :-) The idea looks good, the implementation, I
havent looked in too much details yet but it looks good, though I don't
like the naming of "platform_data" that much (a pain to type, and I find
it confusing to haev a struct, a field and a function all with the same
name).
Ben.
> Signed-off-by: Geoff Levand <geoffrey.levand at am.sony.com>
>
> ---
>
> Index: cell--common--6/arch/powerpc/platforms/cell/setup.c
> ===================================================================
> --- cell--common--6.orig/arch/powerpc/platforms/cell/setup.c
> +++ cell--common--6/arch/powerpc/platforms/cell/setup.c
> @@ -97,7 +97,8 @@
> static void __init cell_setup_arch(void)
> {
> #ifdef CONFIG_SPU_BASE
> - spu_priv1_ops = &spu_priv1_mmio_ops;
> + spu_priv1_ops = &spu_priv1_mmio_ops;
> + spu_management_ops = &spu_management_of_ops;
> #endif
>
> cbe_regs_init();
> Index: cell--common--6/arch/powerpc/platforms/cell/spu_base.c
> ===================================================================
> --- cell--common--6.orig/arch/powerpc/platforms/cell/spu_base.c
> +++ cell--common--6/arch/powerpc/platforms/cell/spu_base.c
> @@ -25,23 +25,17 @@
> #include <linux/interrupt.h>
> #include <linux/list.h>
> #include <linux/module.h>
> -#include <linux/pci.h>
> -#include <linux/poll.h>
> #include <linux/ptrace.h>
> #include <linux/slab.h>
> #include <linux/wait.h>
> -
> -#include <asm/firmware.h>
> -#include <asm/io.h>
> -#include <asm/prom.h>
> +#include <linux/mm.h>
> +#include <linux/io.h>
> #include <linux/mutex.h>
> #include <asm/spu.h>
> #include <asm/spu_priv1.h>
> -#include <asm/mmu_context.h>
> #include <asm/xmon.h>
>
> -#include "interrupt.h"
> -
> +const struct spu_management_ops *spu_management_ops;
> const struct spu_priv1_ops *spu_priv1_ops;
>
> EXPORT_SYMBOL_GPL(spu_priv1_ops);
> @@ -512,235 +506,6 @@
> return ret;
> }
>
> -static int __init find_spu_node_id(struct device_node *spe)
> -{
> - const unsigned int *id;
> - struct device_node *cpu;
> - cpu = spe->parent->parent;
> - id = get_property(cpu, "node-id", NULL);
> - return id ? *id : 0;
> -}
> -
> -static int __init cell_spuprop_present(struct spu *spu, struct device_node *spe,
> - const char *prop)
> -{
> - static DEFINE_MUTEX(add_spumem_mutex);
> -
> - const struct address_prop {
> - unsigned long address;
> - unsigned int len;
> - } __attribute__((packed)) *p;
> - int proplen;
> -
> - unsigned long start_pfn, nr_pages;
> - struct pglist_data *pgdata;
> - struct zone *zone;
> - int ret;
> -
> - p = get_property(spe, prop, &proplen);
> - WARN_ON(proplen != sizeof (*p));
> -
> - start_pfn = p->address >> PAGE_SHIFT;
> - nr_pages = ((unsigned long)p->len + PAGE_SIZE - 1) >> PAGE_SHIFT;
> -
> - pgdata = NODE_DATA(spu->nid);
> - zone = pgdata->node_zones;
> -
> - /* XXX rethink locking here */
> - mutex_lock(&add_spumem_mutex);
> - ret = __add_pages(zone, start_pfn, nr_pages);
> - mutex_unlock(&add_spumem_mutex);
> -
> - return ret;
> -}
> -
> -static void __iomem * __init map_spe_prop(struct spu *spu,
> - struct device_node *n, const char *name)
> -{
> - const struct address_prop {
> - unsigned long address;
> - unsigned int len;
> - } __attribute__((packed)) *prop;
> -
> - const void *p;
> - int proplen;
> - void __iomem *ret = NULL;
> - int err = 0;
> -
> - p = get_property(n, name, &proplen);
> - if (proplen != sizeof (struct address_prop))
> - return NULL;
> -
> - prop = p;
> -
> - err = cell_spuprop_present(spu, n, name);
> - if (err && (err != -EEXIST))
> - goto out;
> -
> - ret = ioremap(prop->address, prop->len);
> -
> - out:
> - return ret;
> -}
> -
> -static void spu_unmap(struct spu *spu)
> -{
> - iounmap(spu->priv2);
> - iounmap(spu->priv1);
> - iounmap(spu->problem);
> - iounmap((__force u8 __iomem *)spu->local_store);
> -}
> -
> -/* This function shall be abstracted for HV platforms */
> -static int __init spu_map_interrupts_old(struct spu *spu, struct device_node *np)
> -{
> - unsigned int isrc;
> - const u32 *tmp;
> -
> - /* Get the interrupt source unit from the device-tree */
> - tmp = get_property(np, "isrc", NULL);
> - if (!tmp)
> - return -ENODEV;
> - isrc = tmp[0];
> -
> - /* Add the node number */
> - isrc |= spu->node << IIC_IRQ_NODE_SHIFT;
> -
> - /* Now map interrupts of all 3 classes */
> - spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
> - spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
> - spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);
> -
> - /* Right now, we only fail if class 2 failed */
> - return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
> -}
> -
> -static int __init spu_map_device_old(struct spu *spu, struct device_node *node)
> -{
> - const char *prop;
> - int ret;
> -
> - ret = -ENODEV;
> - spu->name = get_property(node, "name", NULL);
> - if (!spu->name)
> - goto out;
> -
> - prop = get_property(node, "local-store", NULL);
> - if (!prop)
> - goto out;
> - spu->local_store_phys = *(unsigned long *)prop;
> -
> - /* we use local store as ram, not io memory */
> - spu->local_store = (void __force *)
> - map_spe_prop(spu, node, "local-store");
> - if (!spu->local_store)
> - goto out;
> -
> - prop = get_property(node, "problem", NULL);
> - if (!prop)
> - goto out_unmap;
> - spu->problem_phys = *(unsigned long *)prop;
> -
> - spu->problem= map_spe_prop(spu, node, "problem");
> - if (!spu->problem)
> - goto out_unmap;
> -
> - spu->priv1= map_spe_prop(spu, node, "priv1");
> - /* priv1 is not available on a hypervisor */
> -
> - spu->priv2= map_spe_prop(spu, node, "priv2");
> - if (!spu->priv2)
> - goto out_unmap;
> - ret = 0;
> - goto out;
> -
> -out_unmap:
> - spu_unmap(spu);
> -out:
> - return ret;
> -}
> -
> -static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
> -{
> - struct of_irq oirq;
> - int ret;
> - int i;
> -
> - for (i=0; i < 3; i++) {
> - ret = of_irq_map_one(np, i, &oirq);
> - if (ret)
> - goto err;
> -
> - ret = -EINVAL;
> - spu->irqs[i] = irq_create_of_mapping(oirq.controller,
> - oirq.specifier, oirq.size);
> - if (spu->irqs[i] == NO_IRQ)
> - goto err;
> - }
> - return 0;
> -
> -err:
> - pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier, spu->name);
> - for (; i >= 0; i--) {
> - if (spu->irqs[i] != NO_IRQ)
> - irq_dispose_mapping(spu->irqs[i]);
> - }
> - return ret;
> -}
> -
> -static int spu_map_resource(struct device_node *node, int nr,
> - void __iomem** virt, unsigned long *phys)
> -{
> - struct resource resource = { };
> - int ret;
> -
> - ret = of_address_to_resource(node, 0, &resource);
> - if (ret)
> - goto out;
> -
> - if (phys)
> - *phys = resource.start;
> - *virt = ioremap(resource.start, resource.end - resource.start);
> - if (!*virt)
> - ret = -EINVAL;
> -
> -out:
> - return ret;
> -}
> -
> -static int __init spu_map_device(struct spu *spu, struct device_node *node)
> -{
> - int ret = -ENODEV;
> - spu->name = get_property(node, "name", NULL);
> - if (!spu->name)
> - goto out;
> -
> - ret = spu_map_resource(node, 0, (void __iomem**)&spu->local_store,
> - &spu->local_store_phys);
> - if (ret)
> - goto out;
> - ret = spu_map_resource(node, 1, (void __iomem**)&spu->problem,
> - &spu->problem_phys);
> - if (ret)
> - goto out_unmap;
> - ret = spu_map_resource(node, 2, (void __iomem**)&spu->priv2,
> - NULL);
> - if (ret)
> - goto out_unmap;
> -
> - if (!firmware_has_feature(FW_FEATURE_LPAR))
> - ret = spu_map_resource(node, 3, (void __iomem**)&spu->priv1,
> - NULL);
> - if (ret)
> - goto out_unmap;
> - return 0;
> -
> -out_unmap:
> - spu_unmap(spu);
> -out:
> - pr_debug("failed to map spe %s: %d\n", spu->name, ret);
> - return ret;
> -}
>
> struct sysdev_class spu_sysdev_class = {
> set_kset_name("spu")
> @@ -821,7 +586,7 @@
> sysdev_unregister(&spu->sysdev);
> }
>
> -static int __init create_spu(struct device_node *spe)
> +static int __init create_spu(void *data)
> {
> struct spu *spu;
> int ret;
> @@ -832,60 +597,37 @@
> if (!spu)
> goto out;
>
> - spu->node = find_spu_node_id(spe);
> - if (spu->node >= MAX_NUMNODES) {
> - printk(KERN_WARNING "SPE %s on node %d ignored,"
> - " node number too big\n", spe->full_name, spu->node);
> - printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
> - return -ENODEV;
> - }
> - spu->nid = of_node_to_nid(spe);
> - if (spu->nid == -1)
> - spu->nid = 0;
> + spin_lock_init(&spu->register_lock);
> + mutex_lock(&spu_mutex);
> + spu->number = number++;
> + mutex_unlock(&spu_mutex);
> +
> + ret = spu_create_spu(spu, data);
>
> - ret = spu_map_device(spu, spe);
> - /* try old method */
> - if (ret)
> - ret = spu_map_device_old(spu, spe);
> if (ret)
> goto out_free;
>
> - ret = spu_map_interrupts(spu, spe);
> - if (ret)
> - ret = spu_map_interrupts_old(spu, spe);
> - if (ret)
> - goto out_unmap;
> - spin_lock_init(&spu->register_lock);
> spu_mfc_sdr_setup(spu);
> spu_mfc_sr1_set(spu, 0x33);
> - mutex_lock(&spu_mutex);
> -
> - spu->number = number++;
> ret = spu_request_irqs(spu);
> if (ret)
> - goto out_unlock;
> + goto out_destroy;
>
> ret = spu_create_sysdev(spu);
> if (ret)
> goto out_free_irqs;
>
> + mutex_lock(&spu_mutex);
> list_add(&spu->list, &spu_list[spu->node]);
> list_add(&spu->full_list, &spu_full_list);
> - spu->devnode = of_node_get(spe);
> -
> mutex_unlock(&spu_mutex);
>
> - pr_debug(KERN_DEBUG "Using SPE %s %p %p %p %p %d\n",
> - spu->name, spu->local_store,
> - spu->problem, spu->priv1, spu->priv2, spu->number);
> goto out;
>
> out_free_irqs:
> spu_free_irqs(spu);
> -out_unlock:
> - mutex_unlock(&spu_mutex);
> -out_unmap:
> - spu_unmap(spu);
> +out_destroy:
> + spu_destroy_spu(spu);
> out_free:
> kfree(spu);
> out:
> @@ -897,11 +639,9 @@
> list_del_init(&spu->list);
> list_del_init(&spu->full_list);
>
> - of_node_put(spu->devnode);
> -
> spu_destroy_sysdev(spu);
> spu_free_irqs(spu);
> - spu_unmap(spu);
> + spu_destroy_spu(spu);
> kfree(spu);
> }
>
> @@ -922,7 +662,6 @@
>
> static int __init init_spu_base(void)
> {
> - struct device_node *node;
> int i, ret;
>
> /* create sysdev class for spus */
> @@ -933,16 +672,13 @@
> for (i = 0; i < MAX_NUMNODES; i++)
> INIT_LIST_HEAD(&spu_list[i]);
>
> - ret = -ENODEV;
> - for (node = of_find_node_by_type(NULL, "spe");
> - node; node = of_find_node_by_type(node, "spe")) {
> - ret = create_spu(node);
> - if (ret) {
> - printk(KERN_WARNING "%s: Error initializing %s\n",
> - __FUNCTION__, node->name);
> - cleanup_spu_base();
> - break;
> - }
> + ret = spu_enumerate_spus(create_spu);
> +
> + if (ret) {
> + printk(KERN_WARNING "%s: Error initializing spus\n",
> + __FUNCTION__);
> + cleanup_spu_base();
> + return ret;
> }
>
> xmon_register_spus(&spu_full_list);
> Index: cell--common--6/arch/powerpc/platforms/cell/spu_priv1_mmio.c
> ===================================================================
> --- cell--common--6.orig/arch/powerpc/platforms/cell/spu_priv1_mmio.c
> +++ cell--common--6/arch/powerpc/platforms/cell/spu_priv1_mmio.c
> @@ -18,120 +18,467 @@
> * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
> */
>
> +#undef DEBUG
> +
> +#include <linux/interrupt.h>
> +#include <linux/list.h>
> #include <linux/module.h>
> +#include <linux/ptrace.h>
> +#include <linux/slab.h>
> +#include <linux/wait.h>
> +#include <linux/mm.h>
> +#include <linux/io.h>
> +#include <linux/mutex.h>
> +#include <linux/device.h>
>
> -#include <asm/io.h>
> #include <asm/spu.h>
> #include <asm/spu_priv1.h>
> +#include <asm/firmware.h>
> +#include <asm/prom.h>
>
> #include "interrupt.h"
>
> +struct platform_data {
> + int nid;
> + struct device_node *devnode;
> + struct spu_priv1 __iomem *priv1;
> +};
> +
> +static struct platform_data *platform_data(struct spu *spu)
> +{
> + BUG_ON(!spu->platform_data);
> + return (struct platform_data*)spu->platform_data;
> +}
> +
> +static int __init find_spu_node_id(struct device_node *spe)
> +{
> + const unsigned int *id;
> + struct device_node *cpu;
> + cpu = spe->parent->parent;
> + id = get_property(cpu, "node-id", NULL);
> + return id ? *id : 0;
> +}
> +
> +static int __init cell_spuprop_present(struct spu *spu, struct device_node *spe,
> + const char *prop)
> +{
> + static DEFINE_MUTEX(add_spumem_mutex);
> +
> + const struct address_prop {
> + unsigned long address;
> + unsigned int len;
> + } __attribute__((packed)) *p;
> + int proplen;
> +
> + unsigned long start_pfn, nr_pages;
> + struct pglist_data *pgdata;
> + struct zone *zone;
> + int ret;
> +
> + p = get_property(spe, prop, &proplen);
> + WARN_ON(proplen != sizeof (*p));
> +
> + start_pfn = p->address >> PAGE_SHIFT;
> + nr_pages = ((unsigned long)p->len + PAGE_SIZE - 1) >> PAGE_SHIFT;
> +
> + pgdata = NODE_DATA(spu->nid);
> + zone = pgdata->node_zones;
> +
> + /* XXX rethink locking here */
> + mutex_lock(&add_spumem_mutex);
> + ret = __add_pages(zone, start_pfn, nr_pages);
> + mutex_unlock(&add_spumem_mutex);
> +
> + return ret;
> +}
> +
> +static void __iomem * __init map_spe_prop(struct spu *spu,
> + struct device_node *n, const char *name)
> +{
> + const struct address_prop {
> + unsigned long address;
> + unsigned int len;
> + } __attribute__((packed)) *prop;
> +
> + const void *p;
> + int proplen;
> + void __iomem *ret = NULL;
> + int err = 0;
> +
> + p = get_property(n, name, &proplen);
> + if (proplen != sizeof (struct address_prop))
> + return NULL;
> +
> + prop = p;
> +
> + err = cell_spuprop_present(spu, n, name);
> + if (err && (err != -EEXIST))
> + goto out;
> +
> + ret = ioremap(prop->address, prop->len);
> +
> + out:
> + return ret;
> +}
> +
> +static void spu_unmap(struct spu *spu)
> +{
> + iounmap(spu->priv2);
> + iounmap(platform_data(spu)->priv1);
> + iounmap(spu->problem);
> + iounmap((__force u8 __iomem *)spu->local_store);
> +}
> +
> +static int __init spu_map_interrupts_old(struct spu *spu,
> + struct device_node *np)
> +{
> + unsigned int isrc;
> + const u32 *tmp;
> +
> + /* Get the interrupt source unit from the device-tree */
> + tmp = get_property(np, "isrc", NULL);
> + if (!tmp)
> + return -ENODEV;
> + isrc = tmp[0];
> +
> + /* Add the node number */
> + isrc |= spu->node << IIC_IRQ_NODE_SHIFT;
> +
> + /* Now map interrupts of all 3 classes */
> + spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
> + spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
> + spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);
> +
> + /* Right now, we only fail if class 2 failed */
> + return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
> +}
> +
> +static int __init spu_map_device_old(struct spu *spu, struct device_node *node)
> +{
> + const char *prop;
> + int ret;
> +
> + ret = -ENODEV;
> + spu->name = get_property(node, "name", NULL);
> + if (!spu->name)
> + goto out;
> +
> + prop = get_property(node, "local-store", NULL);
> + if (!prop)
> + goto out;
> + spu->local_store_phys = *(unsigned long *)prop;
> +
> + /* we use local store as ram, not io memory */
> + spu->local_store = (void __force *)
> + map_spe_prop(spu, node, "local-store");
> + if (!spu->local_store)
> + goto out;
> +
> + prop = get_property(node, "problem", NULL);
> + if (!prop)
> + goto out_unmap;
> + spu->problem_phys = *(unsigned long *)prop;
> +
> + spu->problem= map_spe_prop(spu, node, "problem");
> + if (!spu->problem)
> + goto out_unmap;
> +
> + platform_data(spu)->priv1= map_spe_prop(spu, node, "priv1");
> +
> + spu->priv2= map_spe_prop(spu, node, "priv2");
> + if (!spu->priv2)
> + goto out_unmap;
> + ret = 0;
> + goto out;
> +
> +out_unmap:
> + spu_unmap(spu);
> +out:
> + return ret;
> +}
> +
> +static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
> +{
> + struct of_irq oirq;
> + int ret;
> + int i;
> +
> + for (i=0; i < 3; i++) {
> + ret = of_irq_map_one(np, i, &oirq);
> + if (ret)
> + goto err;
> +
> + ret = -EINVAL;
> + spu->irqs[i] = irq_create_of_mapping(oirq.controller,
> + oirq.specifier, oirq.size);
> + if (spu->irqs[i] == NO_IRQ)
> + goto err;
> + }
> + return 0;
> +
> +err:
> + pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier,
> + spu->name);
> + for (; i >= 0; i--) {
> + if (spu->irqs[i] != NO_IRQ)
> + irq_dispose_mapping(spu->irqs[i]);
> + }
> + return ret;
> +}
> +
> +static int spu_map_resource(struct device_node *node, int nr,
> + void __iomem** virt, unsigned long *phys)
> +{
> + struct resource resource = { };
> + int ret;
> +
> + ret = of_address_to_resource(node, 0, &resource);
> + if (ret)
> + goto out;
> +
> + if (phys)
> + *phys = resource.start;
> + *virt = ioremap(resource.start, resource.end - resource.start);
> + if (!*virt)
> + ret = -EINVAL;
> +
> +out:
> + return ret;
> +}
> +
> +static int __init spu_map_device(struct spu *spu, struct device_node *node)
> +{
> + int ret = -ENODEV;
> + spu->name = get_property(node, "name", NULL);
> + if (!spu->name)
> + goto out;
> +
> + ret = spu_map_resource(node, 0, (void __iomem**)&spu->local_store,
> + &spu->local_store_phys);
> + if (ret)
> + goto out;
> + ret = spu_map_resource(node, 1, (void __iomem**)&spu->problem,
> + &spu->problem_phys);
> + if (ret)
> + goto out_unmap;
> + ret = spu_map_resource(node, 2, (void __iomem**)&spu->priv2,
> + NULL);
> + if (ret)
> + goto out_unmap;
> +
> + if (!firmware_has_feature(FW_FEATURE_LPAR))
> + ret = spu_map_resource(node, 3,
> + (void __iomem**)&platform_data(spu)->priv1, NULL);
> + if (ret)
> + goto out_unmap;
> + return 0;
> +
> +out_unmap:
> + spu_unmap(spu);
> +out:
> + pr_debug("failed to map spe %s: %d\n", spu->name, ret);
> + return ret;
> +}
> +
> +static int __init of_create_spu(struct spu *spu, void *data)
> +{
> + int ret;
> + struct device_node *spe = (struct device_node *)data;
> +
> + spu->platform_data = kzalloc(sizeof(struct platform_data),
> + GFP_KERNEL);
> + if (!spu->platform_data) {
> + ret = -ENOMEM;
> + goto out;
> + }
> +
> + spu->node = find_spu_node_id(spe);
> + if (spu->node >= MAX_NUMNODES) {
> + printk(KERN_WARNING "SPE %s on node %d ignored,"
> + " node number too big\n", spe->full_name, spu->node);
> + printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
> + ret = -ENODEV;
> + goto out_free;
> + }
> +
> + platform_data(spu)->nid = of_node_to_nid(spe);
> + if (platform_data(spu)->nid == -1)
> + platform_data(spu)->nid = 0;
> +
> + ret = spu_map_device(spu, spe);
> + /* try old method */
> + if (ret)
> + ret = spu_map_device_old(spu, spe);
> + if (ret)
> + goto out_free;
> +
> + ret = spu_map_interrupts(spu, spe);
> + if (ret)
> + ret = spu_map_interrupts_old(spu, spe);
> + if (ret)
> + goto out_unmap;
> +
> + platform_data(spu)->devnode = of_node_get(spe);
> +
> + pr_debug(KERN_DEBUG "Using SPE %s %p %p %p %p %d\n", spu->name,
> + spu->local_store, spu->problem, platform_data(spu)->priv1,
> + spu->priv2, spu->number);
> + goto out;
> +
> +out_unmap:
> + spu_unmap(spu);
> +out_free:
> + kfree(spu->platform_data);
> + spu->platform_data = NULL;
> +out:
> + return ret;
> +}
> +
> +static int of_destroy_spu(struct spu *spu)
> +{
> + spu_unmap(spu);
> + of_node_put(platform_data(spu)->devnode);
> + kfree(spu->platform_data);
> + spu->platform_data = NULL;
> + return 0;
> +}
> +
> +static int __init of_enumerate_spus(int (*fn)(void *data))
> +{
> + int ret;
> + struct device_node *node;
> +
> + ret = -ENODEV;
> + for (node = of_find_node_by_type(NULL, "spe");
> + node; node = of_find_node_by_type(node, "spe")) {
> + ret = fn(node);
> + if (ret) {
> + printk(KERN_WARNING "%s: Error initializing %s\n",
> + __FUNCTION__, node->name);
> + break;
> + }
> + }
> + return ret;
> +}
> +
> +const struct spu_management_ops spu_management_of_ops = {
> + .enumerate_spus = of_enumerate_spus,
> + .create_spu = of_create_spu,
> + .destroy_spu = of_destroy_spu,
> +};
> +
> static void int_mask_and(struct spu *spu, int class, u64 mask)
> {
> u64 old_mask;
>
> - old_mask = in_be64(&spu->priv1->int_mask_RW[class]);
> - out_be64(&spu->priv1->int_mask_RW[class], old_mask & mask);
> + old_mask = in_be64(&platform_data(spu)->priv1->int_mask_RW[class]);
> + out_be64(&platform_data(spu)->priv1->int_mask_RW[class],
> + old_mask & mask);
> }
>
> static void int_mask_or(struct spu *spu, int class, u64 mask)
> {
> u64 old_mask;
>
> - old_mask = in_be64(&spu->priv1->int_mask_RW[class]);
> - out_be64(&spu->priv1->int_mask_RW[class], old_mask | mask);
> + old_mask = in_be64(&platform_data(spu)->priv1->int_mask_RW[class]);
> + out_be64(&platform_data(spu)->priv1->int_mask_RW[class],
> + old_mask | mask);
> }
>
> static void int_mask_set(struct spu *spu, int class, u64 mask)
> {
> - out_be64(&spu->priv1->int_mask_RW[class], mask);
> + out_be64(&platform_data(spu)->priv1->int_mask_RW[class], mask);
> }
>
> static u64 int_mask_get(struct spu *spu, int class)
> {
> - return in_be64(&spu->priv1->int_mask_RW[class]);
> + return in_be64(&platform_data(spu)->priv1->int_mask_RW[class]);
> }
>
> static void int_stat_clear(struct spu *spu, int class, u64 stat)
> {
> - out_be64(&spu->priv1->int_stat_RW[class], stat);
> + out_be64(&platform_data(spu)->priv1->int_stat_RW[class], stat);
> }
>
> static u64 int_stat_get(struct spu *spu, int class)
> {
> - return in_be64(&spu->priv1->int_stat_RW[class]);
> + return in_be64(&platform_data(spu)->priv1->int_stat_RW[class]);
> }
>
> static void cpu_affinity_set(struct spu *spu, int cpu)
> {
> u64 target = iic_get_target_id(cpu);
> u64 route = target << 48 | target << 32 | target << 16;
> - out_be64(&spu->priv1->int_route_RW, route);
> + out_be64(&platform_data(spu)->priv1->int_route_RW, route);
> }
>
> static u64 mfc_dar_get(struct spu *spu)
> {
> - return in_be64(&spu->priv1->mfc_dar_RW);
> + return in_be64(&platform_data(spu)->priv1->mfc_dar_RW);
> }
>
> static u64 mfc_dsisr_get(struct spu *spu)
> {
> - return in_be64(&spu->priv1->mfc_dsisr_RW);
> + return in_be64(&platform_data(spu)->priv1->mfc_dsisr_RW);
> }
>
> static void mfc_dsisr_set(struct spu *spu, u64 dsisr)
> {
> - out_be64(&spu->priv1->mfc_dsisr_RW, dsisr);
> + out_be64(&platform_data(spu)->priv1->mfc_dsisr_RW, dsisr);
> }
>
> static void mfc_sdr_setup(struct spu *spu)
> {
> - out_be64(&spu->priv1->mfc_sdr_RW, mfspr(SPRN_SDR1));
> + out_be64(&platform_data(spu)->priv1->mfc_sdr_RW, mfspr(SPRN_SDR1));
> }
>
> static void mfc_sr1_set(struct spu *spu, u64 sr1)
> {
> - out_be64(&spu->priv1->mfc_sr1_RW, sr1);
> + out_be64(&platform_data(spu)->priv1->mfc_sr1_RW, sr1);
> }
>
> static u64 mfc_sr1_get(struct spu *spu)
> {
> - return in_be64(&spu->priv1->mfc_sr1_RW);
> + return in_be64(&platform_data(spu)->priv1->mfc_sr1_RW);
> }
>
> static void mfc_tclass_id_set(struct spu *spu, u64 tclass_id)
> {
> - out_be64(&spu->priv1->mfc_tclass_id_RW, tclass_id);
> + out_be64(&platform_data(spu)->priv1->mfc_tclass_id_RW, tclass_id);
> }
>
> static u64 mfc_tclass_id_get(struct spu *spu)
> {
> - return in_be64(&spu->priv1->mfc_tclass_id_RW);
> + return in_be64(&platform_data(spu)->priv1->mfc_tclass_id_RW);
> }
>
> static void tlb_invalidate(struct spu *spu)
> {
> - out_be64(&spu->priv1->tlb_invalidate_entry_W, 0ul);
> + out_be64(&platform_data(spu)->priv1->tlb_invalidate_entry_W, 0ul);
> }
>
> static void resource_allocation_groupID_set(struct spu *spu, u64 id)
> {
> - out_be64(&spu->priv1->resource_allocation_groupID_RW, id);
> + out_be64(&platform_data(spu)->priv1->resource_allocation_groupID_RW,
> + id);
> }
>
> static u64 resource_allocation_groupID_get(struct spu *spu)
> {
> - return in_be64(&spu->priv1->resource_allocation_groupID_RW);
> + return in_be64(
> + &platform_data(spu)->priv1->resource_allocation_groupID_RW);
> }
>
> static void resource_allocation_enable_set(struct spu *spu, u64 enable)
> {
> - out_be64(&spu->priv1->resource_allocation_enable_RW, enable);
> + out_be64(&platform_data(spu)->priv1->resource_allocation_enable_RW,
> + enable);
> }
>
> static u64 resource_allocation_enable_get(struct spu *spu)
> {
> - return in_be64(&spu->priv1->resource_allocation_enable_RW);
> + return in_be64(
> + &platform_data(spu)->priv1->resource_allocation_enable_RW);
> }
>
> const struct spu_priv1_ops spu_priv1_mmio_ops =
> Index: cell--common--6/include/asm-powerpc/spu.h
> ===================================================================
> --- cell--common--6.orig/include/asm-powerpc/spu.h
> +++ cell--common--6/include/asm-powerpc/spu.h
> @@ -111,13 +111,11 @@
> u8 *local_store;
> unsigned long problem_phys;
> struct spu_problem __iomem *problem;
> - struct spu_priv1 __iomem *priv1;
> struct spu_priv2 __iomem *priv2;
> struct list_head list;
> struct list_head sched_list;
> struct list_head full_list;
> int number;
> - int nid;
> unsigned int irqs[3];
> u32 node;
> u64 flags;
> @@ -144,8 +142,7 @@
> char irq_c1[8];
> char irq_c2[8];
>
> - struct device_node *devnode;
> -
> + void* platform_data;
> struct sys_device sysdev;
> };
>
> Index: cell--common--6/include/asm-powerpc/spu_priv1.h
> ===================================================================
> --- cell--common--6.orig/include/asm-powerpc/spu_priv1.h
> +++ cell--common--6/include/asm-powerpc/spu_priv1.h
> @@ -21,12 +21,13 @@
> #define _SPU_PRIV1_H
> #if defined(__KERNEL__)
>
> +#include <linux/types.h>
> +
> struct spu;
>
> /* access to priv1 registers */
>
> -struct spu_priv1_ops
> -{
> +struct spu_priv1_ops {
> void (*int_mask_and) (struct spu *spu, int class, u64 mask);
> void (*int_mask_or) (struct spu *spu, int class, u64 mask);
> void (*int_mask_set) (struct spu *spu, int class, u64 mask);
> @@ -171,12 +172,41 @@
> return spu_priv1_ops->resource_allocation_enable_get(spu);
> }
>
> -/* The declarations folowing are put here for convenience
> - * and only intended to be used by the platform setup code
> - * for initializing spu_priv1_ops.
> +/* spu management abstraction */
> +
> +struct spu_management_ops {
> + int (*enumerate_spus)(int (*fn)(void *data));
> + int (*create_spu)(struct spu *spu, void *data);
> + int (*destroy_spu)(struct spu *spu);
> +};
> +
> +extern const struct spu_management_ops* spu_management_ops;
> +
> +static inline int
> +spu_enumerate_spus (int (*fn)(void *data))
> +{
> + return spu_management_ops->enumerate_spus(fn);
> +}
> +
> +static inline int
> +spu_create_spu (struct spu *spu, void *data)
> +{
> + return spu_management_ops->create_spu(spu, data);
> +}
> +
> +static inline int
> +spu_destroy_spu (struct spu *spu)
> +{
> + return spu_management_ops->destroy_spu(spu);
> +}
> +
> +/*
> + * The declarations folowing are put here for convenience
> + * and only intended to be used by the platform setup code.
> */
>
> extern const struct spu_priv1_ops spu_priv1_mmio_ops;
> +extern const struct spu_management_ops spu_management_of_ops;
>
> #endif /* __KERNEL__ */
> #endif
> _______________________________________________
> Linuxppc-dev mailing list
> Linuxppc-dev at ozlabs.org
> https://ozlabs.org/mailman/listinfo/linuxppc-dev
More information about the cbe-oss-dev
mailing list