[PATCH v4 02/16] powerpc/cell: Move data segment faulting code out of cell platform

Aneesh Kumar K.V aneesh.kumar at linux.vnet.ibm.com
Fri Oct 10 02:04:26 EST 2014


Michael Neuling <mikey at neuling.org> writes:

> From: Ian Munsie <imunsie at au1.ibm.com>
>
> __spu_trap_data_seg() currently contains code to determine the VSID and ESID
> required for a particular EA and mm struct.
>
> This code is generically useful for other co-processors. This moves the code of
> the cell platform so it can be used by other powerpc code. It also adds 1TB
> segment handling which Cell didn't support.  The new function is called
> copro_calculate_slb().
>
> This also moves the internal struct spu_slb to a generic struct copro_slb which
> is now used in the Cell and copro code.  We use this new struct instead of
> passing around esid and vsid parameters.
>

Reviewed-by: Aneesh Kumar K.V <aneesh.kumar at linux.vnet.ibm.com>

> Signed-off-by: Ian Munsie <imunsie at au1.ibm.com>
> Signed-off-by: Michael Neuling <mikey at neuling.org>
> ---
>  arch/powerpc/include/asm/copro.h       |  7 +++++
>  arch/powerpc/include/asm/mmu-hash64.h  |  7 +++++
>  arch/powerpc/mm/copro_fault.c          | 46 ++++++++++++++++++++++++++++
>  arch/powerpc/mm/slb.c                  |  3 --
>  arch/powerpc/platforms/cell/spu_base.c | 55 ++++++----------------------------
>  5 files changed, 69 insertions(+), 49 deletions(-)
>
> diff --git a/arch/powerpc/include/asm/copro.h b/arch/powerpc/include/asm/copro.h
> index 51cae85..b0e6a18 100644
> --- a/arch/powerpc/include/asm/copro.h
> +++ b/arch/powerpc/include/asm/copro.h
> @@ -10,7 +10,14 @@
>  #ifndef _ASM_POWERPC_COPRO_H
>  #define _ASM_POWERPC_COPRO_H
>  
> +struct copro_slb
> +{
> +	u64 esid, vsid;
> +};
> +
>  int copro_handle_mm_fault(struct mm_struct *mm, unsigned long ea,
>  			  unsigned long dsisr, unsigned *flt);
>  
> +int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb);
> +
>  #endif /* _ASM_POWERPC_COPRO_H */
> diff --git a/arch/powerpc/include/asm/mmu-hash64.h b/arch/powerpc/include/asm/mmu-hash64.h
> index d765144..aeabd02 100644
> --- a/arch/powerpc/include/asm/mmu-hash64.h
> +++ b/arch/powerpc/include/asm/mmu-hash64.h
> @@ -190,6 +190,13 @@ static inline unsigned int mmu_psize_to_shift(unsigned int mmu_psize)
>  
>  #ifndef __ASSEMBLY__
>  
> +static inline int slb_vsid_shift(int ssize)
> +{
> +	if (ssize == MMU_SEGSIZE_256M)
> +		return SLB_VSID_SHIFT;
> +	return SLB_VSID_SHIFT_1T;
> +}
> +
>  static inline int segment_shift(int ssize)
>  {
>  	if (ssize == MMU_SEGSIZE_256M)
> diff --git a/arch/powerpc/mm/copro_fault.c b/arch/powerpc/mm/copro_fault.c
> index ba7df14..a15a23e 100644
> --- a/arch/powerpc/mm/copro_fault.c
> +++ b/arch/powerpc/mm/copro_fault.c
> @@ -24,6 +24,7 @@
>  #include <linux/mm.h>
>  #include <linux/export.h>
>  #include <asm/reg.h>
> +#include <asm/copro.h>
>  
>  /*
>   * This ought to be kept in sync with the powerpc specific do_page_fault
> @@ -90,3 +91,48 @@ out_unlock:
>  	return ret;
>  }
>  EXPORT_SYMBOL_GPL(copro_handle_mm_fault);
> +
> +int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb)
> +{
> +	u64 vsid;
> +	int psize, ssize;
> +
> +	slb->esid = (ea & ESID_MASK) | SLB_ESID_V;
> +
> +	switch (REGION_ID(ea)) {
> +	case USER_REGION_ID:
> +		pr_devel("%s: 0x%llx -- USER_REGION_ID\n", __func__, ea);
> +		psize = get_slice_psize(mm, ea);
> +		ssize = user_segment_size(ea);
> +		vsid = get_vsid(mm->context.id, ea, ssize);
> +		break;
> +	case VMALLOC_REGION_ID:
> +		pr_devel("%s: 0x%llx -- VMALLOC_REGION_ID\n", __func__, ea);
> +		if (ea < VMALLOC_END)
> +			psize = mmu_vmalloc_psize;
> +		else
> +			psize = mmu_io_psize;
> +		ssize = mmu_kernel_ssize;
> +		vsid = get_kernel_vsid(ea, mmu_kernel_ssize);
> +		break;
> +	case KERNEL_REGION_ID:
> +		pr_devel("%s: 0x%llx -- KERNEL_REGION_ID\n", __func__, ea);
> +		psize = mmu_linear_psize;
> +		ssize = mmu_kernel_ssize;
> +		vsid = get_kernel_vsid(ea, mmu_kernel_ssize);
> +		break;
> +	default:
> +		pr_debug("%s: invalid region access at %016llx\n", __func__, ea);
> +		return 1;
> +	}
> +
> +	vsid = (vsid << slb_vsid_shift(ssize)) | SLB_VSID_USER;
> +
> +	vsid |= mmu_psize_defs[psize].sllp |
> +		((ssize == MMU_SEGSIZE_1T) ? SLB_VSID_B_1T : 0);
> +

s/0/SLB_VSID_B_256M/ 

> +	slb->vsid = vsid;
> +
> +	return 0;
> +}
> +EXPORT_SYMBOL_GPL(copro_calculate_slb);
> diff --git a/arch/powerpc/mm/slb.c b/arch/powerpc/mm/slb.c
> index 0399a67..6e450ca 100644
> --- a/arch/powerpc/mm/slb.c
> +++ b/arch/powerpc/mm/slb.c
> @@ -46,9 +46,6 @@ static inline unsigned long mk_esid_data(unsigned long ea, int ssize,
>  	return (ea & slb_esid_mask(ssize)) | SLB_ESID_V | slot;
>  }
>  
> -#define slb_vsid_shift(ssize)	\
> -	((ssize) == MMU_SEGSIZE_256M? SLB_VSID_SHIFT: SLB_VSID_SHIFT_1T)
> -
>  static inline unsigned long mk_vsid_data(unsigned long ea, int ssize,
>  					 unsigned long flags)
>  {
> diff --git a/arch/powerpc/platforms/cell/spu_base.c b/arch/powerpc/platforms/cell/spu_base.c
> index 2930d1e..ffcbd24 100644
> --- a/arch/powerpc/platforms/cell/spu_base.c
> +++ b/arch/powerpc/platforms/cell/spu_base.c
> @@ -76,10 +76,6 @@ static LIST_HEAD(spu_full_list);
>  static DEFINE_SPINLOCK(spu_full_list_lock);
>  static DEFINE_MUTEX(spu_full_list_mutex);
>  
> -struct spu_slb {
> -	u64 esid, vsid;
> -};
> -
>  void spu_invalidate_slbs(struct spu *spu)
>  {
>  	struct spu_priv2 __iomem *priv2 = spu->priv2;
> @@ -149,7 +145,7 @@ static void spu_restart_dma(struct spu *spu)
>  	}
>  }
>  
> -static inline void spu_load_slb(struct spu *spu, int slbe, struct spu_slb *slb)
> +static inline void spu_load_slb(struct spu *spu, int slbe, struct copro_slb *slb)
>  {
>  	struct spu_priv2 __iomem *priv2 = spu->priv2;
>  
> @@ -167,45 +163,12 @@ static inline void spu_load_slb(struct spu *spu, int slbe, struct spu_slb *slb)
>  
>  static int __spu_trap_data_seg(struct spu *spu, unsigned long ea)
>  {
> -	struct mm_struct *mm = spu->mm;
> -	struct spu_slb slb;
> -	int psize;
> -
> -	pr_debug("%s\n", __func__);
> -
> -	slb.esid = (ea & ESID_MASK) | SLB_ESID_V;
> +	struct copro_slb slb;
> +	int ret;
>  
> -	switch(REGION_ID(ea)) {
> -	case USER_REGION_ID:
> -#ifdef CONFIG_PPC_MM_SLICES
> -		psize = get_slice_psize(mm, ea);
> -#else
> -		psize = mm->context.user_psize;
> -#endif
> -		slb.vsid = (get_vsid(mm->context.id, ea, MMU_SEGSIZE_256M)
> -				<< SLB_VSID_SHIFT) | SLB_VSID_USER;
> -		break;
> -	case VMALLOC_REGION_ID:
> -		if (ea < VMALLOC_END)
> -			psize = mmu_vmalloc_psize;
> -		else
> -			psize = mmu_io_psize;
> -		slb.vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M)
> -				<< SLB_VSID_SHIFT) | SLB_VSID_KERNEL;
> -		break;
> -	case KERNEL_REGION_ID:
> -		psize = mmu_linear_psize;
> -		slb.vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M)
> -				<< SLB_VSID_SHIFT) | SLB_VSID_KERNEL;
> -		break;
> -	default:
> -		/* Future: support kernel segments so that drivers
> -		 * can use SPUs.
> -		 */
> -		pr_debug("invalid region access at %016lx\n", ea);
> -		return 1;
> -	}
> -	slb.vsid |= mmu_psize_defs[psize].sllp;
> +	ret = copro_calculate_slb(spu->mm, ea, &slb);
> +	if (ret)
> +		return ret;
>  
>  	spu_load_slb(spu, spu->slb_replace, &slb);
>  
> @@ -253,7 +216,7 @@ static int __spu_trap_data_map(struct spu *spu, unsigned long ea, u64 dsisr)
>  	return 0;
>  }
>  
> -static void __spu_kernel_slb(void *addr, struct spu_slb *slb)
> +static void __spu_kernel_slb(void *addr, struct copro_slb *slb)
>  {
>  	unsigned long ea = (unsigned long)addr;
>  	u64 llp;
> @@ -272,7 +235,7 @@ static void __spu_kernel_slb(void *addr, struct spu_slb *slb)
>   * Given an array of @nr_slbs SLB entries, @slbs, return non-zero if the
>   * address @new_addr is present.
>   */
> -static inline int __slb_present(struct spu_slb *slbs, int nr_slbs,
> +static inline int __slb_present(struct copro_slb *slbs, int nr_slbs,
>  		void *new_addr)
>  {
>  	unsigned long ea = (unsigned long)new_addr;
> @@ -297,7 +260,7 @@ static inline int __slb_present(struct spu_slb *slbs, int nr_slbs,
>  void spu_setup_kernel_slbs(struct spu *spu, struct spu_lscsa *lscsa,
>  		void *code, int code_size)
>  {
> -	struct spu_slb slbs[4];
> +	struct copro_slb slbs[4];
>  	int i, nr_slbs = 0;
>  	/* start and end addresses of both mappings */
>  	void *addrs[] = {
> -- 
> 1.9.1
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo at vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/



More information about the Linuxppc-dev mailing list