[PATCH V2] cxl: Export library to support IBM XSL

Frederic Barrat fbarrat at linux.vnet.ibm.com
Thu Jun 15 22:36:24 AEST 2017


Salut Christophe,

A few comments below, nothing major...

Le 14/06/2017 à 15:29, Christophe Lombard a écrit :
> This patch exports a in-kernel 'library' API which can be called by
> other drivers to help interacting with an IBM XSL on a POWER9 system.
> 
> The XSL (Translation Service Layer) is a stripped down version of the
> PSL (Power Service Layer) used in some cards such as the Mellanox CX5.
> Like the PSL, it implements the CAIA architecture, but has a number
> of differences, mostly in it's implementation dependent registers.
> 
> The XSL also uses a special DMA cxl mode, which uses a slightly
> different init sequence for the CAPP and PHB.
> 
> Changelog[v2]
>   - Rebase to latest upstream.
>   - Return -EFAULT in case of NULL pointer in cxllib_handle_fault().
>   - Reverse parameters when copro_handle_mm_fault() is called.
> 


The change log shouldn't be part of the commit message, but below the 
next "---"


> +++ b/drivers/misc/cxl/cxllib.c
> @@ -0,0 +1,241 @@
> +/*
> + * Copyright 2016 IBM Corp.


Year need to be updated


> + *
> + * This program is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU General Public License
> + * as published by the Free Software Foundation; either version
> + * 2 of the License, or (at your option) any later version.
> + */
> +
> +#include <linux/hugetlb.h>
> +#include <linux/sched/mm.h>
> +#include "cxl.h"
> +#include <misc/cxllib.h>
> +#include <asm/pnv-pci.h>

Ordering of the #include is messy:
#include <linux/hugetlb.h>
#include <linux/sched/mm.h>
#include <asm/pnv-pci.h>
#include <misc/cxllib.h>
#include "cxl.h"



> +int cxllib_set_device_dma(struct pci_dev *dev, unsigned long flags)
> +{
> +	int rc;
> +
> +	if (flags)
> +		return -EINVAL;
> +
> +	rc = dma_set_mask(&dev->dev, DMA_BIT_MASK(64));
> +	return rc;
> +}
> +EXPORT_SYMBOL_GPL(cxllib_set_device_dma);


A comment in cxllib_set_device_dma() would help:
	/*
	 * When switching the PHB to capi mode, the TVT#1 entry for
	 * the Partitionable Endpoint is set in bypass mode, like
	 * in PCI mode.
	 * Configure the device dma to use TVT#1, which is done
	 * by calling dma_set_mask() with a mask large enough.
	 */



> +
> +int cxllib_get_PE_attributes(struct task_struct *task,
> +	     unsigned long translation_mode, struct cxllib_pe_attributes *attr)
> +{
> +	struct mm_struct *mm = NULL;
> +
> +	if (translation_mode != CXL_TRANSLATED_MODE &&
> +		translation_mode != CXL_REAL_MODE)
> +		return -EINVAL;
> +
> +	attr->sr = cxl_calculate_sr(false /* master */,
> +				task == NULL /* kernel ctx */,
> +				translation_mode == CXL_REAL_MODE,
> +				true /* p9 */);
> +	attr->lpid = mfspr(SPRN_LPID);
> +	if (task) {
> +		mm = get_task_mm(task);
> +		if (mm == NULL)
> +			return -EINVAL;
> +		/*
> +		 * Caller is keeping a reference on mm_users for as long
> +		 * as XSL uses the memory context
> +		 */
> +		attr->pid = mm->context.id;
> +		mmput(mm);
> +	} else {
> +		attr->pid = 0;
> +	}
> +	attr->tid = 0;


We'll need to remember to patch that function as welll (attr->tid) when 
we add support for as_notify (even though Mellanox is not expected to 
use as_notify in capi mode, just pci I believe)



> +int cxllib_handle_fault(struct mm_struct *mm, u64 addr, u64 size, u64 flags)
> +{
> +	int rc;
> +	u64 dar;
> +	struct vm_area_struct *vma = NULL;
> +	unsigned long page_size;
> +
> +	if (mm == NULL)
> +		return -EFAULT;
> +
> +	down_read(&mm->mmap_sem);
> +
> +	for (dar = addr; dar < addr + size; dar += page_size) {
> +		if (!vma || dar < vma->vm_start || dar > vma->vm_end) {
> +			vma = find_vma(mm, addr);
> +			if (!vma) {
> +				pr_err("Can't find vma for addr %016llx\n", addr);
> +				rc = -EFAULT;
> +				goto out;
> +			}
> +			/* get the size of the pages allocated */
> +			page_size = vma_kernel_pagesize(vma);
> +		}
> +
> +		rc = cxl_handle_page_fault(true, mm, flags, dar);


Why do we pass "true" for kernel parameter?
Actually do we even need a kernel input parameter for 
cxl_handle_page_fault() ? It seems that we can infer it based on mm. If 
NULL, then we are in kernel space.


> diff --git a/drivers/misc/cxl/fault.c b/drivers/misc/cxl/fault.c
> index c79e39b..9db63f3 100644
> --- a/drivers/misc/cxl/fault.c
> +++ b/drivers/misc/cxl/fault.c
> @@ -132,18 +132,16 @@ static int cxl_handle_segment_miss(struct cxl_context *ctx,
>   	return IRQ_HANDLED;
>   }
> 
> -static void cxl_handle_page_fault(struct cxl_context *ctx,
> -				  struct mm_struct *mm, u64 dsisr, u64 dar)
> +int cxl_handle_page_fault(bool kernel_context,
> +			  struct mm_struct *mm, u64 dsisr, u64 dar)
>   {
>   	unsigned flt = 0;
>   	int result;
>   	unsigned long access, flags, inv_flags = 0;
> 
> -	trace_cxl_pte_miss(ctx, dsisr, dar);
> -
>   	if ((result = copro_handle_mm_fault(mm, dar, dsisr, &flt))) {
>   		pr_devel("copro_handle_mm_fault failed: %#x\n", result);
> -		return cxl_ack_ae(ctx);
> +		return result;
>   	}
> 
>   	if (!radix_enabled()) {
> @@ -156,7 +154,7 @@ static void cxl_handle_page_fault(struct cxl_context *ctx,
>   			access |= _PAGE_WRITE;
> 
>   		access |= _PAGE_PRIVILEGED;
> -		if ((!ctx->kernel) || (REGION_ID(dar) == USER_REGION_ID))
> +		if (!kernel_context || (REGION_ID(dar) == USER_REGION_ID))
>   			access &= ~_PAGE_PRIVILEGED;
> 
>   		if (dsisr & DSISR_NOHPTE)
> @@ -166,8 +164,7 @@ static void cxl_handle_page_fault(struct cxl_context *ctx,
>   		hash_page_mm(mm, dar, access, 0x300, inv_flags);
>   		local_irq_restore(flags);
>   	}
> -	pr_devel("Page fault successfully handled for pe: %i!\n", ctx->pe);
> -	cxl_ops->ack_irq(ctx, CXL_PSL_TFC_An_R, 0);
> +	return 0;
>   }
> 
>   /*
> @@ -261,9 +258,15 @@ void cxl_handle_fault(struct work_struct *fault_work)
> 
>   	if (cxl_is_segment_miss(ctx, dsisr))
>   		cxl_handle_segment_miss(ctx, mm, dar);
> -	else if (cxl_is_page_fault(ctx, dsisr))
> -		cxl_handle_page_fault(ctx, mm, dsisr, dar);
> -	else
> +	else if (cxl_is_page_fault(ctx, dsisr)) {
> +		trace_cxl_pte_miss(ctx, dsisr, dar);
> +		if (cxl_handle_page_fault(ctx->kernel, mm, dsisr, dar)) {
> +			cxl_ack_ae(ctx);
> +		} else {
> +			pr_devel("Page fault successfully handled for pe: %i!\n", ctx->pe);
> +			cxl_ops->ack_irq(ctx, CXL_PSL_TFC_An_R, 0);
> +		}


Could we have that code in a wrapper before calling 
cxl_handle_page_fault()? It would keep the code cleaner and in line with 
what we do for cxl_handle_segment_miss().



> +++ b/include/misc/cxllib.h
> @@ -0,0 +1,132 @@
> +/*
> + * Copyright 2016 IBM Corp.


Year update.


> +/*
> + * Get the Process Element structure for the given thread
> + *
> + * Input:
> + *	pid: points the struct pid for the given thread (i.e. linux pid)
> + *	translation_mode: whether addresses should be translated
> + */
> +struct cxllib_pe_attributes {
> +	u64 sr;
> +	u32 lpid;
> +	u32 tid;
> +	u32 pid;
> +};
> +#define CXL_TRANSLATED_MODE 0
> +#define CXL_REAL_MODE 1
> +
> +int cxllib_get_PE_attributes(struct task_struct *task,
> +	     unsigned long translation_mode, struct cxllib_pe_attributes *attr);


Description in comment no longer matches reality.
/*
  * Get the Process Element structure for the given thread
  *
  * Input:
  *	task: task_struct for the context of the translation
  *	translation_mode: whether addresses should be translated
  * Output:
  *	attr: attributes to fill up the Process Element structure from CAIA
  */


Thanks,

   Fred



More information about the Linuxppc-dev mailing list