[Cbe-oss-dev] [PATCH] cell: abstract spu management routines

Benjamin Herrenschmidt benh at kernel.crashing.org
Wed Nov 8 16:31:35 EST 2006


> For the IBM Cell Blade support, I put the hypervisor only resources that were
> in struct spu into a platform specific data structure struct platform_data.

We have no hypervisor :-) The idea looks good, the implementation, I
havent looked in too much details yet but it looks good, though I don't
like the naming of "platform_data" that much (a pain to type, and I find
it confusing to haev a struct, a field and a function all with the same
name).

Ben.


> Signed-off-by: Geoff Levand <geoffrey.levand at am.sony.com>
> 
> ---
> 
> Index: cell--common--6/arch/powerpc/platforms/cell/setup.c
> ===================================================================
> --- cell--common--6.orig/arch/powerpc/platforms/cell/setup.c
> +++ cell--common--6/arch/powerpc/platforms/cell/setup.c
> @@ -97,7 +97,8 @@
>  static void __init cell_setup_arch(void)
>  {
>  #ifdef CONFIG_SPU_BASE
> -	spu_priv1_ops         = &spu_priv1_mmio_ops;
> +	spu_priv1_ops = &spu_priv1_mmio_ops;
> +	spu_management_ops = &spu_management_of_ops;
>  #endif
>  
>  	cbe_regs_init();
> Index: cell--common--6/arch/powerpc/platforms/cell/spu_base.c
> ===================================================================
> --- cell--common--6.orig/arch/powerpc/platforms/cell/spu_base.c
> +++ cell--common--6/arch/powerpc/platforms/cell/spu_base.c
> @@ -25,23 +25,17 @@
>  #include <linux/interrupt.h>
>  #include <linux/list.h>
>  #include <linux/module.h>
> -#include <linux/pci.h>
> -#include <linux/poll.h>
>  #include <linux/ptrace.h>
>  #include <linux/slab.h>
>  #include <linux/wait.h>
> -
> -#include <asm/firmware.h>
> -#include <asm/io.h>
> -#include <asm/prom.h>
> +#include <linux/mm.h>
> +#include <linux/io.h>
>  #include <linux/mutex.h>
>  #include <asm/spu.h>
>  #include <asm/spu_priv1.h>
> -#include <asm/mmu_context.h>
>  #include <asm/xmon.h>
>  
> -#include "interrupt.h"
> -
> +const struct spu_management_ops *spu_management_ops;
>  const struct spu_priv1_ops *spu_priv1_ops;
>  
>  EXPORT_SYMBOL_GPL(spu_priv1_ops);
> @@ -512,235 +506,6 @@
>  	return ret;
>  }
>  
> -static int __init find_spu_node_id(struct device_node *spe)
> -{
> -	const unsigned int *id;
> -	struct device_node *cpu;
> -	cpu = spe->parent->parent;
> -	id = get_property(cpu, "node-id", NULL);
> -	return id ? *id : 0;
> -}
> -
> -static int __init cell_spuprop_present(struct spu *spu, struct device_node *spe,
> -		const char *prop)
> -{
> -	static DEFINE_MUTEX(add_spumem_mutex);
> -
> -	const struct address_prop {
> -		unsigned long address;
> -		unsigned int len;
> -	} __attribute__((packed)) *p;
> -	int proplen;
> -
> -	unsigned long start_pfn, nr_pages;
> -	struct pglist_data *pgdata;
> -	struct zone *zone;
> -	int ret;
> -
> -	p = get_property(spe, prop, &proplen);
> -	WARN_ON(proplen != sizeof (*p));
> -
> -	start_pfn = p->address >> PAGE_SHIFT;
> -	nr_pages = ((unsigned long)p->len + PAGE_SIZE - 1) >> PAGE_SHIFT;
> -
> -	pgdata = NODE_DATA(spu->nid);
> -	zone = pgdata->node_zones;
> -
> -	/* XXX rethink locking here */
> -	mutex_lock(&add_spumem_mutex);
> -	ret = __add_pages(zone, start_pfn, nr_pages);
> -	mutex_unlock(&add_spumem_mutex);
> -
> -	return ret;
> -}
> -
> -static void __iomem * __init map_spe_prop(struct spu *spu,
> -		struct device_node *n, const char *name)
> -{
> -	const struct address_prop {
> -		unsigned long address;
> -		unsigned int len;
> -	} __attribute__((packed)) *prop;
> -
> -	const void *p;
> -	int proplen;
> -	void __iomem *ret = NULL;
> -	int err = 0;
> -
> -	p = get_property(n, name, &proplen);
> -	if (proplen != sizeof (struct address_prop))
> -		return NULL;
> -
> -	prop = p;
> -
> -	err = cell_spuprop_present(spu, n, name);
> -	if (err && (err != -EEXIST))
> -		goto out;
> -
> -	ret = ioremap(prop->address, prop->len);
> -
> - out:
> -	return ret;
> -}
> -
> -static void spu_unmap(struct spu *spu)
> -{
> -	iounmap(spu->priv2);
> -	iounmap(spu->priv1);
> -	iounmap(spu->problem);
> -	iounmap((__force u8 __iomem *)spu->local_store);
> -}
> -
> -/* This function shall be abstracted for HV platforms */
> -static int __init spu_map_interrupts_old(struct spu *spu, struct device_node *np)
> -{
> -	unsigned int isrc;
> -	const u32 *tmp;
> -
> -	/* Get the interrupt source unit from the device-tree */
> -	tmp = get_property(np, "isrc", NULL);
> -	if (!tmp)
> -		return -ENODEV;
> -	isrc = tmp[0];
> -
> -	/* Add the node number */
> -	isrc |= spu->node << IIC_IRQ_NODE_SHIFT;
> -
> -	/* Now map interrupts of all 3 classes */
> -	spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
> -	spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
> -	spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);
> -
> -	/* Right now, we only fail if class 2 failed */
> -	return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
> -}
> -
> -static int __init spu_map_device_old(struct spu *spu, struct device_node *node)
> -{
> -	const char *prop;
> -	int ret;
> -
> -	ret = -ENODEV;
> -	spu->name = get_property(node, "name", NULL);
> -	if (!spu->name)
> -		goto out;
> -
> -	prop = get_property(node, "local-store", NULL);
> -	if (!prop)
> -		goto out;
> -	spu->local_store_phys = *(unsigned long *)prop;
> -
> -	/* we use local store as ram, not io memory */
> -	spu->local_store = (void __force *)
> -		map_spe_prop(spu, node, "local-store");
> -	if (!spu->local_store)
> -		goto out;
> -
> -	prop = get_property(node, "problem", NULL);
> -	if (!prop)
> -		goto out_unmap;
> -	spu->problem_phys = *(unsigned long *)prop;
> -
> -	spu->problem= map_spe_prop(spu, node, "problem");
> -	if (!spu->problem)
> -		goto out_unmap;
> -
> -	spu->priv1= map_spe_prop(spu, node, "priv1");
> -	/* priv1 is not available on a hypervisor */
> -
> -	spu->priv2= map_spe_prop(spu, node, "priv2");
> -	if (!spu->priv2)
> -		goto out_unmap;
> -	ret = 0;
> -	goto out;
> -
> -out_unmap:
> -	spu_unmap(spu);
> -out:
> -	return ret;
> -}
> -
> -static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
> -{
> -	struct of_irq oirq;
> -	int ret;
> -	int i;
> -
> -	for (i=0; i < 3; i++) {
> -		ret = of_irq_map_one(np, i, &oirq);
> -		if (ret)
> -			goto err;
> -
> -		ret = -EINVAL;
> -		spu->irqs[i] = irq_create_of_mapping(oirq.controller,
> -					oirq.specifier, oirq.size);
> -		if (spu->irqs[i] == NO_IRQ)
> -			goto err;
> -	}
> -	return 0;
> -
> -err:
> -	pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier, spu->name);
> -	for (; i >= 0; i--) {
> -		if (spu->irqs[i] != NO_IRQ)
> -			irq_dispose_mapping(spu->irqs[i]);
> -	}
> -	return ret;
> -}
> -
> -static int spu_map_resource(struct device_node *node, int nr,
> -		void __iomem** virt, unsigned long *phys)
> -{
> -	struct resource resource = { };
> -	int ret;
> -
> -	ret = of_address_to_resource(node, 0, &resource);
> -	if (ret)
> -		goto out;
> -
> -	if (phys)
> -		*phys = resource.start;
> -	*virt = ioremap(resource.start, resource.end - resource.start);
> -	if (!*virt)
> -		ret = -EINVAL;
> -
> -out:
> -	return ret;
> -}
> -
> -static int __init spu_map_device(struct spu *spu, struct device_node *node)
> -{
> -	int ret = -ENODEV;
> -	spu->name = get_property(node, "name", NULL);
> -	if (!spu->name)
> -		goto out;
> -
> -	ret = spu_map_resource(node, 0, (void __iomem**)&spu->local_store,
> -					&spu->local_store_phys);
> -	if (ret)
> -		goto out;
> -	ret = spu_map_resource(node, 1, (void __iomem**)&spu->problem,
> -					&spu->problem_phys);
> -	if (ret)
> -		goto out_unmap;
> -	ret = spu_map_resource(node, 2, (void __iomem**)&spu->priv2,
> -					NULL);
> -	if (ret)
> -		goto out_unmap;
> -
> -	if (!firmware_has_feature(FW_FEATURE_LPAR))
> -		ret = spu_map_resource(node, 3, (void __iomem**)&spu->priv1,
> -					NULL);
> -	if (ret)
> -		goto out_unmap;
> -	return 0;
> -
> -out_unmap:
> -	spu_unmap(spu);
> -out:
> -	pr_debug("failed to map spe %s: %d\n", spu->name, ret);
> -	return ret;
> -}
>  
>  struct sysdev_class spu_sysdev_class = {
>  	set_kset_name("spu")
> @@ -821,7 +586,7 @@
>  	sysdev_unregister(&spu->sysdev);
>  }
>  
> -static int __init create_spu(struct device_node *spe)
> +static int __init create_spu(void *data)
>  {
>  	struct spu *spu;
>  	int ret;
> @@ -832,60 +597,37 @@
>  	if (!spu)
>  		goto out;
>  
> -	spu->node = find_spu_node_id(spe);
> -	if (spu->node >= MAX_NUMNODES) {
> -		printk(KERN_WARNING "SPE %s on node %d ignored,"
> -		       " node number too big\n", spe->full_name, spu->node);
> -		printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
> -		return -ENODEV;
> -	}
> -	spu->nid = of_node_to_nid(spe);
> -	if (spu->nid == -1)
> -		spu->nid = 0;
> +	spin_lock_init(&spu->register_lock);
> +	mutex_lock(&spu_mutex);
> +	spu->number = number++;
> +	mutex_unlock(&spu_mutex);
> +
> +	ret = spu_create_spu(spu, data);
>  
> -	ret = spu_map_device(spu, spe);
> -	/* try old method */
> -	if (ret)
> -		ret = spu_map_device_old(spu, spe);
>  	if (ret)
>  		goto out_free;
>  
> -	ret = spu_map_interrupts(spu, spe);
> -	if (ret)
> -		ret = spu_map_interrupts_old(spu, spe);
> -	if (ret)
> -		goto out_unmap;
> -	spin_lock_init(&spu->register_lock);
>  	spu_mfc_sdr_setup(spu);
>  	spu_mfc_sr1_set(spu, 0x33);
> -	mutex_lock(&spu_mutex);
> -
> -	spu->number = number++;
>  	ret = spu_request_irqs(spu);
>  	if (ret)
> -		goto out_unlock;
> +		goto out_destroy;
>  
>  	ret = spu_create_sysdev(spu);
>  	if (ret)
>  		goto out_free_irqs;
>  
> +	mutex_lock(&spu_mutex);
>  	list_add(&spu->list, &spu_list[spu->node]);
>  	list_add(&spu->full_list, &spu_full_list);
> -	spu->devnode = of_node_get(spe);
> -
>  	mutex_unlock(&spu_mutex);
>  
> -	pr_debug(KERN_DEBUG "Using SPE %s %p %p %p %p %d\n",
> -		spu->name, spu->local_store,
> -		spu->problem, spu->priv1, spu->priv2, spu->number);
>  	goto out;
>  
>  out_free_irqs:
>  	spu_free_irqs(spu);
> -out_unlock:
> -	mutex_unlock(&spu_mutex);
> -out_unmap:
> -	spu_unmap(spu);
> +out_destroy:
> +	spu_destroy_spu(spu);
>  out_free:
>  	kfree(spu);
>  out:
> @@ -897,11 +639,9 @@
>  	list_del_init(&spu->list);
>  	list_del_init(&spu->full_list);
>  
> -	of_node_put(spu->devnode);
> -
>  	spu_destroy_sysdev(spu);
>  	spu_free_irqs(spu);
> -	spu_unmap(spu);
> +	spu_destroy_spu(spu);
>  	kfree(spu);
>  }
>  
> @@ -922,7 +662,6 @@
>  
>  static int __init init_spu_base(void)
>  {
> -	struct device_node *node;
>  	int i, ret;
>  
>  	/* create sysdev class for spus */
> @@ -933,16 +672,13 @@
>  	for (i = 0; i < MAX_NUMNODES; i++)
>  		INIT_LIST_HEAD(&spu_list[i]);
>  
> -	ret = -ENODEV;
> -	for (node = of_find_node_by_type(NULL, "spe");
> -			node; node = of_find_node_by_type(node, "spe")) {
> -		ret = create_spu(node);
> -		if (ret) {
> -			printk(KERN_WARNING "%s: Error initializing %s\n",
> -				__FUNCTION__, node->name);
> -			cleanup_spu_base();
> -			break;
> -		}
> +	ret = spu_enumerate_spus(create_spu);
> +
> +	if (ret) {
> +		printk(KERN_WARNING "%s: Error initializing spus\n",
> +			__FUNCTION__);
> +		cleanup_spu_base();
> +		return ret;
>  	}
>  
>  	xmon_register_spus(&spu_full_list);
> Index: cell--common--6/arch/powerpc/platforms/cell/spu_priv1_mmio.c
> ===================================================================
> --- cell--common--6.orig/arch/powerpc/platforms/cell/spu_priv1_mmio.c
> +++ cell--common--6/arch/powerpc/platforms/cell/spu_priv1_mmio.c
> @@ -18,120 +18,467 @@
>   *  Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
>   */
>  
> +#undef DEBUG
> +
> +#include <linux/interrupt.h>
> +#include <linux/list.h>
>  #include <linux/module.h>
> +#include <linux/ptrace.h>
> +#include <linux/slab.h>
> +#include <linux/wait.h>
> +#include <linux/mm.h>
> +#include <linux/io.h>
> +#include <linux/mutex.h>
> +#include <linux/device.h>
>  
> -#include <asm/io.h>
>  #include <asm/spu.h>
>  #include <asm/spu_priv1.h>
> +#include <asm/firmware.h>
> +#include <asm/prom.h>
>  
>  #include "interrupt.h"
>  
> +struct platform_data {
> +	int nid;
> +	struct device_node *devnode;
> +	struct spu_priv1 __iomem *priv1;
> +};
> +
> +static struct platform_data *platform_data(struct spu *spu)
> +{
> +	BUG_ON(!spu->platform_data);
> +	return (struct platform_data*)spu->platform_data;
> +}
> +
> +static int __init find_spu_node_id(struct device_node *spe)
> +{
> +	const unsigned int *id;
> +	struct device_node *cpu;
> +	cpu = spe->parent->parent;
> +	id = get_property(cpu, "node-id", NULL);
> +	return id ? *id : 0;
> +}
> +
> +static int __init cell_spuprop_present(struct spu *spu, struct device_node *spe,
> +		const char *prop)
> +{
> +	static DEFINE_MUTEX(add_spumem_mutex);
> +
> +	const struct address_prop {
> +		unsigned long address;
> +		unsigned int len;
> +	} __attribute__((packed)) *p;
> +	int proplen;
> +
> +	unsigned long start_pfn, nr_pages;
> +	struct pglist_data *pgdata;
> +	struct zone *zone;
> +	int ret;
> +
> +	p = get_property(spe, prop, &proplen);
> +	WARN_ON(proplen != sizeof (*p));
> +
> +	start_pfn = p->address >> PAGE_SHIFT;
> +	nr_pages = ((unsigned long)p->len + PAGE_SIZE - 1) >> PAGE_SHIFT;
> +
> +	pgdata = NODE_DATA(spu->nid);
> +	zone = pgdata->node_zones;
> +
> +	/* XXX rethink locking here */
> +	mutex_lock(&add_spumem_mutex);
> +	ret = __add_pages(zone, start_pfn, nr_pages);
> +	mutex_unlock(&add_spumem_mutex);
> +
> +	return ret;
> +}
> +
> +static void __iomem * __init map_spe_prop(struct spu *spu,
> +		struct device_node *n, const char *name)
> +{
> +	const struct address_prop {
> +		unsigned long address;
> +		unsigned int len;
> +	} __attribute__((packed)) *prop;
> +
> +	const void *p;
> +	int proplen;
> +	void __iomem *ret = NULL;
> +	int err = 0;
> +
> +	p = get_property(n, name, &proplen);
> +	if (proplen != sizeof (struct address_prop))
> +		return NULL;
> +
> +	prop = p;
> +
> +	err = cell_spuprop_present(spu, n, name);
> +	if (err && (err != -EEXIST))
> +		goto out;
> +
> +	ret = ioremap(prop->address, prop->len);
> +
> + out:
> +	return ret;
> +}
> +
> +static void spu_unmap(struct spu *spu)
> +{
> +	iounmap(spu->priv2);
> +	iounmap(platform_data(spu)->priv1);
> +	iounmap(spu->problem);
> +	iounmap((__force u8 __iomem *)spu->local_store);
> +}
> +
> +static int __init spu_map_interrupts_old(struct spu *spu,
> +	struct device_node *np)
> +{
> +	unsigned int isrc;
> +	const u32 *tmp;
> +
> +	/* Get the interrupt source unit from the device-tree */
> +	tmp = get_property(np, "isrc", NULL);
> +	if (!tmp)
> +		return -ENODEV;
> +	isrc = tmp[0];
> +
> +	/* Add the node number */
> +	isrc |= spu->node << IIC_IRQ_NODE_SHIFT;
> +
> +	/* Now map interrupts of all 3 classes */
> +	spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
> +	spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
> +	spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);
> +
> +	/* Right now, we only fail if class 2 failed */
> +	return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
> +}
> +
> +static int __init spu_map_device_old(struct spu *spu, struct device_node *node)
> +{
> +	const char *prop;
> +	int ret;
> +
> +	ret = -ENODEV;
> +	spu->name = get_property(node, "name", NULL);
> +	if (!spu->name)
> +		goto out;
> +
> +	prop = get_property(node, "local-store", NULL);
> +	if (!prop)
> +		goto out;
> +	spu->local_store_phys = *(unsigned long *)prop;
> +
> +	/* we use local store as ram, not io memory */
> +	spu->local_store = (void __force *)
> +		map_spe_prop(spu, node, "local-store");
> +	if (!spu->local_store)
> +		goto out;
> +
> +	prop = get_property(node, "problem", NULL);
> +	if (!prop)
> +		goto out_unmap;
> +	spu->problem_phys = *(unsigned long *)prop;
> +
> +	spu->problem= map_spe_prop(spu, node, "problem");
> +	if (!spu->problem)
> +		goto out_unmap;
> +
> +	platform_data(spu)->priv1= map_spe_prop(spu, node, "priv1");
> +
> +	spu->priv2= map_spe_prop(spu, node, "priv2");
> +	if (!spu->priv2)
> +		goto out_unmap;
> +	ret = 0;
> +	goto out;
> +
> +out_unmap:
> +	spu_unmap(spu);
> +out:
> +	return ret;
> +}
> +
> +static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
> +{
> +	struct of_irq oirq;
> +	int ret;
> +	int i;
> +
> +	for (i=0; i < 3; i++) {
> +		ret = of_irq_map_one(np, i, &oirq);
> +		if (ret)
> +			goto err;
> +
> +		ret = -EINVAL;
> +		spu->irqs[i] = irq_create_of_mapping(oirq.controller,
> +					oirq.specifier, oirq.size);
> +		if (spu->irqs[i] == NO_IRQ)
> +			goto err;
> +	}
> +	return 0;
> +
> +err:
> +	pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier,
> +		spu->name);
> +	for (; i >= 0; i--) {
> +		if (spu->irqs[i] != NO_IRQ)
> +			irq_dispose_mapping(spu->irqs[i]);
> +	}
> +	return ret;
> +}
> +
> +static int spu_map_resource(struct device_node *node, int nr,
> +		void __iomem** virt, unsigned long *phys)
> +{
> +	struct resource resource = { };
> +	int ret;
> +
> +	ret = of_address_to_resource(node, 0, &resource);
> +	if (ret)
> +		goto out;
> +
> +	if (phys)
> +		*phys = resource.start;
> +	*virt = ioremap(resource.start, resource.end - resource.start);
> +	if (!*virt)
> +		ret = -EINVAL;
> +
> +out:
> +	return ret;
> +}
> +
> +static int __init spu_map_device(struct spu *spu, struct device_node *node)
> +{
> +	int ret = -ENODEV;
> +	spu->name = get_property(node, "name", NULL);
> +	if (!spu->name)
> +		goto out;
> +
> +	ret = spu_map_resource(node, 0, (void __iomem**)&spu->local_store,
> +					&spu->local_store_phys);
> +	if (ret)
> +		goto out;
> +	ret = spu_map_resource(node, 1, (void __iomem**)&spu->problem,
> +					&spu->problem_phys);
> +	if (ret)
> +		goto out_unmap;
> +	ret = spu_map_resource(node, 2, (void __iomem**)&spu->priv2,
> +					NULL);
> +	if (ret)
> +		goto out_unmap;
> +
> +	if (!firmware_has_feature(FW_FEATURE_LPAR))
> +		ret = spu_map_resource(node, 3,
> +			(void __iomem**)&platform_data(spu)->priv1, NULL);
> +	if (ret)
> +		goto out_unmap;
> +	return 0;
> +
> +out_unmap:
> +	spu_unmap(spu);
> +out:
> +	pr_debug("failed to map spe %s: %d\n", spu->name, ret);
> +	return ret;
> +}
> +
> +static int __init of_create_spu(struct spu *spu, void *data)
> +{
> +	int ret;
> +	struct device_node *spe = (struct device_node *)data;
> +
> +	spu->platform_data = kzalloc(sizeof(struct platform_data),
> +		GFP_KERNEL);
> +	if (!spu->platform_data) {
> +		ret = -ENOMEM;
> +		goto out;
> +	}
> +
> +	spu->node = find_spu_node_id(spe);
> +	if (spu->node >= MAX_NUMNODES) {
> +		printk(KERN_WARNING "SPE %s on node %d ignored,"
> +		       " node number too big\n", spe->full_name, spu->node);
> +		printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
> +		ret = -ENODEV;
> +		goto out_free;
> +	}
> +
> +	platform_data(spu)->nid = of_node_to_nid(spe);
> +	if (platform_data(spu)->nid == -1)
> +		platform_data(spu)->nid = 0;
> +
> +	ret = spu_map_device(spu, spe);
> +	/* try old method */
> +	if (ret)
> +		ret = spu_map_device_old(spu, spe);
> +	if (ret)
> +		goto out_free;
> +
> +	ret = spu_map_interrupts(spu, spe);
> +	if (ret)
> +		ret = spu_map_interrupts_old(spu, spe);
> +	if (ret)
> +		goto out_unmap;
> +
> +	platform_data(spu)->devnode = of_node_get(spe);
> +
> +	pr_debug(KERN_DEBUG "Using SPE %s %p %p %p %p %d\n", spu->name,
> +		spu->local_store, spu->problem, platform_data(spu)->priv1,
> +		spu->priv2, spu->number);
> +	goto out;
> +
> +out_unmap:
> +	spu_unmap(spu);
> +out_free:
> +	kfree(spu->platform_data);
> +	spu->platform_data = NULL;
> +out:
> +	return ret;
> +}
> +
> +static int of_destroy_spu(struct spu *spu)
> +{
> +	spu_unmap(spu);
> +	of_node_put(platform_data(spu)->devnode);
> +	kfree(spu->platform_data);
> +	spu->platform_data = NULL;
> +	return 0;
> +}
> +
> +static int __init of_enumerate_spus(int (*fn)(void *data))
> +{
> +	int ret;
> +	struct device_node *node;
> +
> +	ret = -ENODEV;
> +	for (node = of_find_node_by_type(NULL, "spe");
> +			node; node = of_find_node_by_type(node, "spe")) {
> +		ret = fn(node);
> +		if (ret) {
> +			printk(KERN_WARNING "%s: Error initializing %s\n",
> +				__FUNCTION__, node->name);
> +			break;
> +		}
> +	}
> +	return ret;
> +}
> +
> +const struct spu_management_ops spu_management_of_ops = {
> +	.enumerate_spus = of_enumerate_spus,
> +	.create_spu = of_create_spu,
> +	.destroy_spu = of_destroy_spu,
> +};
> +
>  static void int_mask_and(struct spu *spu, int class, u64 mask)
>  {
>  	u64 old_mask;
>  
> -	old_mask = in_be64(&spu->priv1->int_mask_RW[class]);
> -	out_be64(&spu->priv1->int_mask_RW[class], old_mask & mask);
> +	old_mask = in_be64(&platform_data(spu)->priv1->int_mask_RW[class]);
> +	out_be64(&platform_data(spu)->priv1->int_mask_RW[class],
> +		old_mask & mask);
>  }
>  
>  static void int_mask_or(struct spu *spu, int class, u64 mask)
>  {
>  	u64 old_mask;
>  
> -	old_mask = in_be64(&spu->priv1->int_mask_RW[class]);
> -	out_be64(&spu->priv1->int_mask_RW[class], old_mask | mask);
> +	old_mask = in_be64(&platform_data(spu)->priv1->int_mask_RW[class]);
> +	out_be64(&platform_data(spu)->priv1->int_mask_RW[class],
> +		old_mask | mask);
>  }
>  
>  static void int_mask_set(struct spu *spu, int class, u64 mask)
>  {
> -	out_be64(&spu->priv1->int_mask_RW[class], mask);
> +	out_be64(&platform_data(spu)->priv1->int_mask_RW[class], mask);
>  }
>  
>  static u64 int_mask_get(struct spu *spu, int class)
>  {
> -	return in_be64(&spu->priv1->int_mask_RW[class]);
> +	return in_be64(&platform_data(spu)->priv1->int_mask_RW[class]);
>  }
>  
>  static void int_stat_clear(struct spu *spu, int class, u64 stat)
>  {
> -	out_be64(&spu->priv1->int_stat_RW[class], stat);
> +	out_be64(&platform_data(spu)->priv1->int_stat_RW[class], stat);
>  }
>  
>  static u64 int_stat_get(struct spu *spu, int class)
>  {
> -	return in_be64(&spu->priv1->int_stat_RW[class]);
> +	return in_be64(&platform_data(spu)->priv1->int_stat_RW[class]);
>  }
>  
>  static void cpu_affinity_set(struct spu *spu, int cpu)
>  {
>  	u64 target = iic_get_target_id(cpu);
>  	u64 route = target << 48 | target << 32 | target << 16;
> -	out_be64(&spu->priv1->int_route_RW, route);
> +	out_be64(&platform_data(spu)->priv1->int_route_RW, route);
>  }
>  
>  static u64 mfc_dar_get(struct spu *spu)
>  {
> -	return in_be64(&spu->priv1->mfc_dar_RW);
> +	return in_be64(&platform_data(spu)->priv1->mfc_dar_RW);
>  }
>  
>  static u64 mfc_dsisr_get(struct spu *spu)
>  {
> -	return in_be64(&spu->priv1->mfc_dsisr_RW);
> +	return in_be64(&platform_data(spu)->priv1->mfc_dsisr_RW);
>  }
>  
>  static void mfc_dsisr_set(struct spu *spu, u64 dsisr)
>  {
> -	out_be64(&spu->priv1->mfc_dsisr_RW, dsisr);
> +	out_be64(&platform_data(spu)->priv1->mfc_dsisr_RW, dsisr);
>  }
>  
>  static void mfc_sdr_setup(struct spu *spu)
>  {
> -	out_be64(&spu->priv1->mfc_sdr_RW, mfspr(SPRN_SDR1));
> +	out_be64(&platform_data(spu)->priv1->mfc_sdr_RW, mfspr(SPRN_SDR1));
>  }
>  
>  static void mfc_sr1_set(struct spu *spu, u64 sr1)
>  {
> -	out_be64(&spu->priv1->mfc_sr1_RW, sr1);
> +	out_be64(&platform_data(spu)->priv1->mfc_sr1_RW, sr1);
>  }
>  
>  static u64 mfc_sr1_get(struct spu *spu)
>  {
> -	return in_be64(&spu->priv1->mfc_sr1_RW);
> +	return in_be64(&platform_data(spu)->priv1->mfc_sr1_RW);
>  }
>  
>  static void mfc_tclass_id_set(struct spu *spu, u64 tclass_id)
>  {
> -	out_be64(&spu->priv1->mfc_tclass_id_RW, tclass_id);
> +	out_be64(&platform_data(spu)->priv1->mfc_tclass_id_RW, tclass_id);
>  }
>  
>  static u64 mfc_tclass_id_get(struct spu *spu)
>  {
> -	return in_be64(&spu->priv1->mfc_tclass_id_RW);
> +	return in_be64(&platform_data(spu)->priv1->mfc_tclass_id_RW);
>  }
>  
>  static void tlb_invalidate(struct spu *spu)
>  {
> -	out_be64(&spu->priv1->tlb_invalidate_entry_W, 0ul);
> +	out_be64(&platform_data(spu)->priv1->tlb_invalidate_entry_W, 0ul);
>  }
>  
>  static void resource_allocation_groupID_set(struct spu *spu, u64 id)
>  {
> -	out_be64(&spu->priv1->resource_allocation_groupID_RW, id);
> +	out_be64(&platform_data(spu)->priv1->resource_allocation_groupID_RW,
> +		id);
>  }
>  
>  static u64 resource_allocation_groupID_get(struct spu *spu)
>  {
> -	return in_be64(&spu->priv1->resource_allocation_groupID_RW);
> +	return in_be64(
> +		&platform_data(spu)->priv1->resource_allocation_groupID_RW);
>  }
>  
>  static void resource_allocation_enable_set(struct spu *spu, u64 enable)
>  {
> -	out_be64(&spu->priv1->resource_allocation_enable_RW, enable);
> +	out_be64(&platform_data(spu)->priv1->resource_allocation_enable_RW,
> +		enable);
>  }
>  
>  static u64 resource_allocation_enable_get(struct spu *spu)
>  {
> -	return in_be64(&spu->priv1->resource_allocation_enable_RW);
> +	return in_be64(
> +		&platform_data(spu)->priv1->resource_allocation_enable_RW);
>  }
>  
>  const struct spu_priv1_ops spu_priv1_mmio_ops =
> Index: cell--common--6/include/asm-powerpc/spu.h
> ===================================================================
> --- cell--common--6.orig/include/asm-powerpc/spu.h
> +++ cell--common--6/include/asm-powerpc/spu.h
> @@ -111,13 +111,11 @@
>  	u8 *local_store;
>  	unsigned long problem_phys;
>  	struct spu_problem __iomem *problem;
> -	struct spu_priv1 __iomem *priv1;
>  	struct spu_priv2 __iomem *priv2;
>  	struct list_head list;
>  	struct list_head sched_list;
>  	struct list_head full_list;
>  	int number;
> -	int nid;
>  	unsigned int irqs[3];
>  	u32 node;
>  	u64 flags;
> @@ -144,8 +142,7 @@
>  	char irq_c1[8];
>  	char irq_c2[8];
>  
> -	struct device_node *devnode;
> -
> +	void* platform_data;
>  	struct sys_device sysdev;
>  };
>  
> Index: cell--common--6/include/asm-powerpc/spu_priv1.h
> ===================================================================
> --- cell--common--6.orig/include/asm-powerpc/spu_priv1.h
> +++ cell--common--6/include/asm-powerpc/spu_priv1.h
> @@ -21,12 +21,13 @@
>  #define _SPU_PRIV1_H
>  #if defined(__KERNEL__)
>  
> +#include <linux/types.h>
> +
>  struct spu;
>  
>  /* access to priv1 registers */
>  
> -struct spu_priv1_ops
> -{
> +struct spu_priv1_ops {
>  	void (*int_mask_and) (struct spu *spu, int class, u64 mask);
>  	void (*int_mask_or) (struct spu *spu, int class, u64 mask);
>  	void (*int_mask_set) (struct spu *spu, int class, u64 mask);
> @@ -171,12 +172,41 @@
>  	return spu_priv1_ops->resource_allocation_enable_get(spu);
>  }
>  
> -/* The declarations folowing are put here for convenience
> - * and only intended to be used by the platform setup code
> - * for initializing spu_priv1_ops.
> +/* spu management abstraction */
> +
> +struct spu_management_ops {
> +	int (*enumerate_spus)(int (*fn)(void *data));
> +	int (*create_spu)(struct spu *spu, void *data);
> +	int (*destroy_spu)(struct spu *spu);
> +};
> +
> +extern const struct spu_management_ops* spu_management_ops;
> +
> +static inline int
> +spu_enumerate_spus (int (*fn)(void *data))
> +{
> +	return spu_management_ops->enumerate_spus(fn);
> +}
> +
> +static inline int
> +spu_create_spu (struct spu *spu, void *data)
> +{
> +	return spu_management_ops->create_spu(spu, data);
> +}
> +
> +static inline int
> +spu_destroy_spu (struct spu *spu)
> +{
> +	return spu_management_ops->destroy_spu(spu);
> +}
> +
> +/*
> + * The declarations folowing are put here for convenience
> + * and only intended to be used by the platform setup code.
>   */
>  
>  extern const struct spu_priv1_ops spu_priv1_mmio_ops;
> +extern const struct spu_management_ops spu_management_of_ops;
>  
>  #endif /* __KERNEL__ */
>  #endif
> _______________________________________________
> Linuxppc-dev mailing list
> Linuxppc-dev at ozlabs.org
> https://ozlabs.org/mailman/listinfo/linuxppc-dev




More information about the cbe-oss-dev mailing list