Fwd: [PATCH v4 12/18] cxl: Add guest-specific code

Manoj Kumar manoj at linux.vnet.ibm.com
Mon Feb 22 12:29:27 AEDT 2016


Christophe, Fred:

Is getting the code checkpatch clean not a requirement for
this component?

total: 458 errors, 995 warnings, 1602 lines checked

NOTE: Whitespace errors detected.
       You may wish to use scripts/cleanpatch or scripts/cleanfile


I am stopping my review at this point.
Will pick it back up after you resubmit.

--
Manoj Kumar

> Subject: [PATCH v4 12/18] cxl: Add guest-specific code
> Date: Tue, 16 Feb 2016 22:39:05 +0100
> From: Frederic Barrat <fbarrat at linux.vnet.ibm.com>
> To: imunsie at au1.ibm.com, michael.neuling at au1.ibm.com,
> mpe at ellerman.id.au, linuxppc-dev at lists.ozlabs.org
>
> From: Christophe Lombard <clombard at linux.vnet.ibm.com>
>
> The new of.c file contains code to parse the device tree to find out
> about CAPI adapters and AFUs.
>
> guest.c implements the guest-specific callbacks for the backend API.
>
> The process element ID is not known until the context is attached, so
> we have to separate the context ID assigned by the cxl driver from the
> process element ID visible to the user applications. In bare-metal,
> the 2 IDs match.
>
> Co-authored-by: Frederic Barrat <fbarrat at linux.vnet.ibm.com>
> Signed-off-by: Frederic Barrat <fbarrat at linux.vnet.ibm.com>
> Signed-off-by: Christophe Lombard <clombard at linux.vnet.ibm.com>
> ---
>   drivers/misc/cxl/Makefile  |   1 +
>   drivers/misc/cxl/api.c     |   2 +-
>   drivers/misc/cxl/context.c |   6 +-
>   drivers/misc/cxl/cxl.h     |  37 +-
>   drivers/misc/cxl/file.c    |   2 +-
>   drivers/misc/cxl/guest.c   | 950
> +++++++++++++++++++++++++++++++++++++++++++++
>   drivers/misc/cxl/main.c    |  18 +-
>   drivers/misc/cxl/of.c      | 513 ++++++++++++++++++++++++
>   8 files changed, 1519 insertions(+), 10 deletions(-)
>   create mode 100644 drivers/misc/cxl/guest.c
>   create mode 100644 drivers/misc/cxl/of.c
>
> diff --git a/drivers/misc/cxl/Makefile b/drivers/misc/cxl/Makefile
> index be2ac5c..a3d4bef 100644
> --- a/drivers/misc/cxl/Makefile
> +++ b/drivers/misc/cxl/Makefile
> @@ -4,6 +4,7 @@ ccflags-$(CONFIG_PPC_WERROR)    += -Werror
>   cxl-y                += main.o file.o irq.o fault.o native.o
>   cxl-y                += context.o sysfs.o debugfs.o pci.o trace.o
>   cxl-y                += vphb.o api.o
> +cxl-y                += guest.o of.o hcalls.o
>   obj-$(CONFIG_CXL)        += cxl.o
>   obj-$(CONFIG_CXL_BASE)        += base.o
>
> diff --git a/drivers/misc/cxl/api.c b/drivers/misc/cxl/api.c
> index 31eb842..325f957 100644
> --- a/drivers/misc/cxl/api.c
> +++ b/drivers/misc/cxl/api.c
> @@ -191,7 +191,7 @@ EXPORT_SYMBOL_GPL(cxl_start_context);
>
>   int cxl_process_element(struct cxl_context *ctx)
>   {
> -    return ctx->pe;
> +    return ctx->external_pe;
>   }
>   EXPORT_SYMBOL_GPL(cxl_process_element);
>
> diff --git a/drivers/misc/cxl/context.c b/drivers/misc/cxl/context.c
> index 200837f..180c85a 100644
> --- a/drivers/misc/cxl/context.c
> +++ b/drivers/misc/cxl/context.c
> @@ -95,8 +95,12 @@ int cxl_context_init(struct cxl_context *ctx, struct
> cxl_afu *afu, bool master,
>           return i;
>
>       ctx->pe = i;
> -    if (cpu_has_feature(CPU_FTR_HVMODE))
> +    if (cpu_has_feature(CPU_FTR_HVMODE)) {
>           ctx->elem = &ctx->afu->native->spa[i];
> +        ctx->external_pe = ctx->pe;
> +    } else {
> +        ctx->external_pe = -1; /* assigned when attaching */
> +    }
>       ctx->pe_inserted = false;
>
>       /*
> diff --git a/drivers/misc/cxl/cxl.h b/drivers/misc/cxl/cxl.h
> index 3a1fabd..4372a87 100644
> --- a/drivers/misc/cxl/cxl.h
> +++ b/drivers/misc/cxl/cxl.h
> @@ -433,6 +433,12 @@ struct cxl_irq_name {
>       char *name;
>   };
>
> +struct irq_avail {
> +    irq_hw_number_t offset;
> +    irq_hw_number_t range;
> +    unsigned long   *bitmap;
> +};
> +
>   /*
>    * This is a cxl context.  If the PSL is in dedicated mode, there will
> be one
>    * of these per AFU.  If in AFU directed there can be lots of these.
> @@ -488,7 +494,19 @@ struct cxl_context {
>
>       struct cxl_process_element *elem;
>
> -    int pe; /* process element handle */
> +    /*
> +     * pe is the process element handle, assigned by this driver when the
> +     * context is initialized.
> +     *
> +     * external_pe is the PE shown outside of cxl.
> +     * On bare-metal, pe=external_pe, because we decide what the handle
> is.
> +     * In a guest, we only find out about the pe used by pHyp when the
> +     * context is attached, and that's the value we want to report outside
> +     * of cxl.
> +     */
> +    int pe;
> +    int external_pe;
> +
>       u32 irq_count;
>       bool pe_inserted;
>       bool master;
> @@ -782,6 +800,7 @@ void cxl_pci_vphb_reconfigure(struct cxl_afu *afu);
>   void cxl_pci_vphb_remove(struct cxl_afu *afu);
>
>   extern struct pci_driver cxl_pci_driver;
> +extern struct platform_driver cxl_of_driver;
>   int afu_allocate_irqs(struct cxl_context *ctx, u32 count);
>
>   int afu_open(struct inode *inode, struct file *file);
> @@ -792,6 +811,21 @@ unsigned int afu_poll(struct file *file, struct
> poll_table_struct *poll);
>   ssize_t afu_read(struct file *file, char __user *buf, size_t count,
> loff_t *off);
>   extern const struct file_operations afu_fops;
>
> +struct cxl *cxl_guest_init_adapter(struct device_node *np, struct
> platform_device *dev);
> +void cxl_guest_remove_adapter(struct cxl *adapter);
> +int cxl_of_read_adapter_handle(struct cxl *adapter, struct device_node
> *np);
> +int cxl_of_read_adapter_properties(struct cxl *adapter, struct
> device_node *np);
> +ssize_t cxl_guest_read_adapter_vpd(struct cxl *adapter, void *buf,
> size_t len);
> +ssize_t cxl_guest_read_afu_vpd(struct cxl_afu *afu, void *buf, size_t
> len);
> +int cxl_guest_init_afu(struct cxl *adapter, int slice, struct
> device_node *afu_np);
> +void cxl_guest_remove_afu(struct cxl_afu *afu);
> +int cxl_of_read_afu_handle(struct cxl_afu *afu, struct device_node
> *afu_np);
> +int cxl_of_read_afu_properties(struct cxl_afu *afu, struct device_node
> *afu_np);
> +int cxl_guest_add_chardev(struct cxl *adapter);
> +void cxl_guest_remove_chardev(struct cxl *adapter);
> +void cxl_guest_reload_module(struct cxl *adapter);
> +int cxl_of_probe(struct platform_device *pdev);
> +
>   struct cxl_backend_ops {
>       struct module *module;
>       int (*adapter_reset)(struct cxl *adapter);
> @@ -824,6 +858,7 @@ struct cxl_backend_ops {
>       int (*afu_cr_read64)(struct cxl_afu *afu, int cr_idx, u64 offset,
> u64 *val);
>   };
>   extern const struct cxl_backend_ops cxl_native_ops;
> +extern const struct cxl_backend_ops cxl_guest_ops;
>   extern const struct cxl_backend_ops *cxl_ops;
>
>   #endif
> diff --git a/drivers/misc/cxl/file.c b/drivers/misc/cxl/file.c
> index b8ce29b..df4d49a 100644
> --- a/drivers/misc/cxl/file.c
> +++ b/drivers/misc/cxl/file.c
> @@ -228,7 +228,7 @@ static long afu_ioctl_process_element(struct
> cxl_context *ctx,
>   {
>       pr_devel("%s: pe: %i\n", __func__, ctx->pe);
>
> -    if (copy_to_user(upe, &ctx->pe, sizeof(__u32)))
> +    if (copy_to_user(upe, &ctx->external_pe, sizeof(__u32)))
>           return -EFAULT;
>
>       return 0;
> diff --git a/drivers/misc/cxl/guest.c b/drivers/misc/cxl/guest.c
> new file mode 100644
> index 0000000..03eb83d
> --- /dev/null
> +++ b/drivers/misc/cxl/guest.c
> @@ -0,0 +1,950 @@
> +/*
> + * Copyright 2015 IBM Corp.
> + *
> + * This program is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU General Public License
> + * as published by the Free Software Foundation; either version
> + * 2 of the License, or (at your option) any later version.
> + */
> +
> +#include <linux/spinlock.h>
> +#include <linux/uaccess.h>
> +#include <linux/delay.h>
> +
> +#include "cxl.h"
> +#include "hcalls.h"
> +#include "trace.h"
> +
> +
> +static irqreturn_t guest_handle_psl_slice_error(struct cxl_context
> *ctx, u64 dsisr,
> +                    u64 errstat)
> +{
> +    pr_devel("in %s\n", __func__);
> +    dev_crit(&ctx->afu->dev, "PSL ERROR STATUS: 0x%.16llx\n", errstat);
> +
> +    return cxl_ops->ack_irq(ctx, 0, errstat);
> +}
> +
> +static ssize_t guest_collect_vpd(struct cxl *adapter, struct cxl_afu *afu,
> +            void *buf, size_t len)
> +{
> +    unsigned int entries, mod;
> +    unsigned long **vpd_buf = NULL;
> +    struct sg_list *le;
> +    int rc = 0, i, tocopy;
> +    u64 out = 0;
> +
> +    if (buf == NULL)
> +        return -EINVAL;
> +
> +    /* number of entries in the list */
> +    entries = len / SG_BUFFER_SIZE;
> +    mod = len % SG_BUFFER_SIZE;
> +    if (mod)
> +        entries++;
> +
> +    if (entries > SG_MAX_ENTRIES) {
> +        entries = SG_MAX_ENTRIES;
> +        len = SG_MAX_ENTRIES * SG_BUFFER_SIZE;
> +        mod = 0;
> +    }
> +
> +    vpd_buf = kzalloc(entries * sizeof(unsigned long *), GFP_KERNEL);
> +    if (!vpd_buf)
> +        return -ENOMEM;
> +
> +    le = (struct sg_list *)get_zeroed_page(GFP_KERNEL);
> +    if (!le) {
> +        rc = -ENOMEM;
> +        goto err1;
> +    }
> +
> +    for (i = 0; i < entries; i++) {
> +        vpd_buf[i] = (unsigned long *)get_zeroed_page(GFP_KERNEL);
> +        if (!vpd_buf[i]) {
> +            rc = -ENOMEM;
> +            goto err2;
> +        }
> +        le[i].phys_addr = cpu_to_be64(virt_to_phys(vpd_buf[i]));
> +        le[i].len = cpu_to_be64(SG_BUFFER_SIZE);
> +        if ((i == (entries - 1)) && mod)
> +            le[i].len = cpu_to_be64(mod);
> +    }
> +
> +    if (adapter)
> +        rc = cxl_h_collect_vpd_adapter(adapter->guest->handle,
> +                    virt_to_phys(le), entries, &out);
> +    else
> +        rc = cxl_h_collect_vpd(afu->guest->handle, 0,
> +                virt_to_phys(le), entries, &out);
> +    pr_devel("length of available (entries: %i), vpd: %#llx\n",
> +        entries, out);
> +
> +    if (!rc) {
> +        /*
> +         * hcall returns in 'out' the size of available VPDs.
> +         * It fills the buffer with as much data as possible.
> +         */
> +        if (out < len)
> +            len = out;
> +        rc = len;
> +        if (out) {
> +            for (i = 0; i < entries; i++) {
> +                if (len < SG_BUFFER_SIZE)
> +                    tocopy = len;
> +                else
> +                    tocopy = SG_BUFFER_SIZE;
> +                memcpy(buf, vpd_buf[i], tocopy);
> +                buf += tocopy;
> +                len -= tocopy;
> +            }
> +        }
> +    }
> +err2:
> +    for (i = 0; i < entries; i++) {
> +        if (vpd_buf[i])
> +            free_page((unsigned long) vpd_buf[i]);
> +    }
> +    free_page((unsigned long) le);
> +err1:
> +    kfree(vpd_buf);
> +    return rc;
> +}
> +
> +static int guest_get_irq_info(struct cxl_context *ctx, struct
> cxl_irq_info *info)
> +{
> +    return cxl_h_collect_int_info(ctx->afu->guest->handle,
> ctx->process_token, info);
> +}
> +
> +static irqreturn_t guest_psl_irq(int irq, void *data)
> +{
> +    struct cxl_context *ctx = data;
> +    struct cxl_irq_info irq_info;
> +    int rc;
> +
> +    pr_devel("%d: received PSL interrupt %i\n", ctx->pe, irq);
> +    rc = guest_get_irq_info(ctx, &irq_info);
> +    if (rc) {
> +        WARN(1, "Unable to get IRQ info: %i\n", rc);
> +        return IRQ_HANDLED;
> +    }
> +
> +    rc = cxl_irq(irq, ctx, &irq_info);
> +    return rc;
> +}
> +
> +static irqreturn_t guest_slice_irq_err(int irq, void *data)
> +{
> +    struct cxl_afu *afu = data;
> +    int rc;
> +    u64 serr;
> +
> +    WARN(irq, "CXL SLICE ERROR interrupt %i\n", irq);
> +    rc = cxl_h_get_fn_error_interrupt(afu->guest->handle, &serr);
> +    if (rc) {
> +        dev_crit(&afu->dev, "Couldn't read PSL_SERR_An: %d\n", rc);
> +        return IRQ_HANDLED;
> +    }
> +    dev_crit(&afu->dev, "PSL_SERR_An: 0x%.16llx\n", serr);
> +
> +    rc = cxl_h_ack_fn_error_interrupt(afu->guest->handle, serr);
> +    if (rc)
> +        dev_crit(&afu->dev, "Couldn't ack slice error interrupt: %d\n",
> +            rc);
> +
> +    return IRQ_HANDLED;
> +}
> +
> +
> +static int irq_alloc_range(struct cxl *adapter, int len, int *irq)
> +{
> +    int i, n;
> +    struct irq_avail *cur;
> +
> +    for (i = 0; i < adapter->guest->irq_nranges; i++) {
> +        cur = &adapter->guest->irq_avail[i];
> +        n = bitmap_find_next_zero_area(cur->bitmap, cur->range,
> +                    0, len, 0);
> +        if (n < cur->range) {
> +            bitmap_set(cur->bitmap, n, len);
> +            *irq = cur->offset + n;
> +            pr_devel("guest: allocate IRQs %#x->%#x\n",
> +                *irq, *irq + len - 1);
> +
> +            return 0;
> +        }
> +    }
> +    return -ENOSPC;
> +}
> +
> +static int irq_free_range(struct cxl *adapter, int irq, int len)
> +{
> +    int i, n;
> +    struct irq_avail *cur;
> +
> +    if (len == 0)
> +        return -ENOENT;
> +
> +    for (i = 0; i < adapter->guest->irq_nranges; i++) {
> +        cur = &adapter->guest->irq_avail[i];
> +        if (irq >= cur->offset &&
> +            (irq + len) <= (cur->offset + cur->range)) {
> +            n = irq - cur->offset;
> +            bitmap_clear(cur->bitmap, n, len);
> +            pr_devel("guest: release IRQs %#x->%#x\n",
> +                irq, irq + len - 1);
> +            return 0;
> +        }
> +    }
> +    return -ENOENT;
> +}
> +
> +static int guest_reset(struct cxl *adapter)
> +{
> +    int rc;
> +
> +    pr_devel("Adapter reset request\n");
> +    rc = cxl_h_reset_adapter(adapter->guest->handle);
> +    return rc;
> +}
> +
> +static int guest_alloc_one_irq(struct cxl *adapter)
> +{
> +    int irq;
> +
> +    spin_lock(&adapter->guest->irq_alloc_lock);
> +    if (irq_alloc_range(adapter, 1, &irq))
> +        irq = -ENOSPC;
> +    spin_unlock(&adapter->guest->irq_alloc_lock);
> +    return irq;
> +}
> +
> +static void guest_release_one_irq(struct cxl *adapter, int irq)
> +{
> +    spin_lock(&adapter->guest->irq_alloc_lock);
> +    irq_free_range(adapter, irq, 1);
> +    spin_unlock(&adapter->guest->irq_alloc_lock);
> +}
> +
> +static int guest_alloc_irq_ranges(struct cxl_irq_ranges *irqs,
> +                struct cxl *adapter, unsigned int num)
> +{
> +    int i, try, irq;
> +
> +    memset(irqs, 0, sizeof(struct cxl_irq_ranges));
> +
> +    spin_lock(&adapter->guest->irq_alloc_lock);
> +    for (i = 0; i < CXL_IRQ_RANGES && num; i++) {
> +        try = num;
> +        while (try) {
> +            if (irq_alloc_range(adapter, try, &irq) == 0)
> +                break;
> +            try /= 2;
> +        }
> +        if (!try)
> +            goto error;
> +        irqs->offset[i] = irq;
> +        irqs->range[i] = try;
> +        num -= try;
> +    }
> +    if (num)
> +        goto error;
> +    spin_unlock(&adapter->guest->irq_alloc_lock);
> +    return 0;
> +
> +error:
> +    for (i = 0; i < CXL_IRQ_RANGES; i++)
> +        irq_free_range(adapter, irqs->offset[i], irqs->range[i]);
> +    spin_unlock(&adapter->guest->irq_alloc_lock);
> +    return -ENOSPC;
> +}
> +
> +static void guest_release_irq_ranges(struct cxl_irq_ranges *irqs,
> +                struct cxl *adapter)
> +{
> +    int i;
> +
> +    spin_lock(&adapter->guest->irq_alloc_lock);
> +    for (i = 0; i < CXL_IRQ_RANGES; i++)
> +        irq_free_range(adapter, irqs->offset[i], irqs->range[i]);
> +    spin_unlock(&adapter->guest->irq_alloc_lock);
> +}
> +
> +static int guest_register_serr_irq(struct cxl_afu *afu)
> +{
> +    afu->err_irq_name = kasprintf(GFP_KERNEL, "cxl-%s-err",
> +                      dev_name(&afu->dev));
> +    if (!afu->err_irq_name)
> +        return -ENOMEM;
> +
> +    if (!(afu->serr_virq = cxl_map_irq(afu->adapter, afu->serr_hwirq,
> +                 guest_slice_irq_err, afu, afu->err_irq_name))) {
> +        kfree(afu->err_irq_name);
> +        afu->err_irq_name = NULL;
> +        return -ENOMEM;
> +    }
> +
> +    return 0;
> +}
> +
> +static void guest_release_serr_irq(struct cxl_afu *afu)
> +{
> +    cxl_unmap_irq(afu->serr_virq, afu);
> +    cxl_ops->release_one_irq(afu->adapter, afu->serr_hwirq);
> +    kfree(afu->err_irq_name);
> +}
> +
> +static int guest_ack_irq(struct cxl_context *ctx, u64 tfc, u64
> psl_reset_mask)
> +{
> +    return cxl_h_control_faults(ctx->afu->guest->handle,
> ctx->process_token,
> +                tfc >> 32, (psl_reset_mask != 0));
> +}
> +
> +static void disable_afu_irqs(struct cxl_context *ctx)
> +{
> +    irq_hw_number_t hwirq;
> +    unsigned int virq;
> +    int r, i;
> +
> +    pr_devel("Disabling AFU(%d) interrupts\n", ctx->afu->slice);
> +    for (r = 0; r < CXL_IRQ_RANGES; r++) {
> +        hwirq = ctx->irqs.offset[r];
> +        for (i = 0; i < ctx->irqs.range[r]; hwirq++, i++) {
> +            virq = irq_find_mapping(NULL, hwirq);
> +            disable_irq(virq);
> +        }
> +    }
> +}
> +
> +static void enable_afu_irqs(struct cxl_context *ctx)
> +{
> +    irq_hw_number_t hwirq;
> +    unsigned int virq;
> +    int r, i;
> +
> +    pr_devel("Enabling AFU(%d) interrupts\n", ctx->afu->slice);
> +    for (r = 0; r < CXL_IRQ_RANGES; r++) {
> +        hwirq = ctx->irqs.offset[r];
> +        for (i = 0; i < ctx->irqs.range[r]; hwirq++, i++) {
> +            virq = irq_find_mapping(NULL, hwirq);
> +            enable_irq(virq);
> +        }
> +    }
> +}
> +
> +static int _guest_afu_cr_readXX(int sz, struct cxl_afu *afu, int cr_idx,
> +            u64 offset, u64 *val)
> +{
> +    unsigned long cr;
> +    char c;
> +    int rc = 0;
> +
> +    if (afu->crs_len < sz)
> +        return -ENOENT;
> +
> +    if (unlikely(offset >= afu->crs_len))
> +        return -ERANGE;
> +
> +    cr = get_zeroed_page(GFP_KERNEL);
> +    if (!cr)
> +        return -ENOMEM;
> +
> +    rc = cxl_h_get_config(afu->guest->handle, cr_idx, offset,
> +            virt_to_phys((void *)cr), sz);
> +    if (rc)
> +        goto err;
> +
> +    switch (sz) {
> +    case 1:
> +        c = *((char *) cr);
> +        *val = c;
> +        break;
> +    case 2:
> +        *val = in_le16((u16 *)cr);
> +        break;
> +    case 4:
> +        *val = in_le32((unsigned *)cr);
> +        break;
> +    case 8:
> +        *val = in_le64((u64 *)cr);
> +        break;
> +    default:
> +        WARN_ON(1);
> +    }
> +err:
> +    free_page(cr);
> +    return rc;
> +}
> +
> +static int guest_afu_cr_read32(struct cxl_afu *afu, int cr_idx,    u64
> offset,
> +            u32 *out)
> +{
> +    int rc;
> +    u64 val;
> +
> +    rc = _guest_afu_cr_readXX(4, afu, cr_idx, offset, &val);
> +    if (!rc)
> +        *out = (u32) val;
> +    return rc;
> +}
> +
> +static int guest_afu_cr_read16(struct cxl_afu *afu, int cr_idx,    u64
> offset,
> +            u16 *out)
> +{
> +    int rc;
> +    u64 val;
> +
> +    rc = _guest_afu_cr_readXX(2, afu, cr_idx, offset, &val);
> +    if (!rc)
> +        *out = (u16) val;
> +    return rc;
> +}
> +
> +static int guest_afu_cr_read8(struct cxl_afu *afu, int cr_idx, u64 offset,
> +            u8 *out)
> +{
> +    int rc;
> +    u64 val;
> +
> +    rc = _guest_afu_cr_readXX(1, afu, cr_idx, offset, &val);
> +    if (!rc)
> +        *out = (u8) val;
> +    return rc;
> +}
> +
> +static int guest_afu_cr_read64(struct cxl_afu *afu, int cr_idx,    u64
> offset,
> +            u64 *out)
> +{
> +    return _guest_afu_cr_readXX(8, afu, cr_idx, offset, out);
> +}
> +
> +static int attach_afu_directed(struct cxl_context *ctx, u64 wed, u64 amr)
> +{
> +    struct cxl_process_element_hcall *elem;
> +    struct cxl *adapter = ctx->afu->adapter;
> +    const struct cred *cred;
> +    u32 pid, idx;
> +    int rc, r, i;
> +    u64 mmio_addr, mmio_size;
> +    __be64 flags = 0;
> +
> +    /* Must be 8 byte aligned and cannot cross a 4096 byte boundary */
> +    if (!(elem = (struct cxl_process_element_hcall *)
> +            get_zeroed_page(GFP_KERNEL)))
> +        return -ENOMEM;
> +
> +    elem->version = cpu_to_be64(CXL_PROCESS_ELEMENT_VERSION);
> +    if (ctx->kernel) {
> +        pid = 0;
> +        flags |= CXL_PE_TRANSLATION_ENABLED;
> +        flags |= CXL_PE_PRIVILEGED_PROCESS;
> +        if (mfmsr() & MSR_SF)
> +            flags |= CXL_PE_64_BIT;
> +    } else {
> +        pid = current->pid;
> +        flags |= CXL_PE_PROBLEM_STATE;
> +        flags |= CXL_PE_TRANSLATION_ENABLED;
> +        if (!test_tsk_thread_flag(current, TIF_32BIT))
> +            flags |= CXL_PE_64_BIT;
> +        cred = get_current_cred();
> +        if (uid_eq(cred->euid, GLOBAL_ROOT_UID))
> +            flags |= CXL_PE_PRIVILEGED_PROCESS;
> +        put_cred(cred);
> +    }
> +    elem->flags         = cpu_to_be64(flags);
> +    elem->common.tid    = cpu_to_be32(0); /* Unused */
> +    elem->common.pid    = cpu_to_be32(pid);
> +    elem->common.csrp   = cpu_to_be64(0); /* disable */
> +    elem->common.aurp0  = cpu_to_be64(0); /* disable */
> +    elem->common.aurp1  = cpu_to_be64(0); /* disable */
> +
> +    cxl_prefault(ctx, wed);
> +
> +    elem->common.sstp0  = cpu_to_be64(ctx->sstp0);
> +    elem->common.sstp1  = cpu_to_be64(ctx->sstp1);
> +    for (r = 0; r < CXL_IRQ_RANGES; r++) {
> +        for (i = 0; i < ctx->irqs.range[r]; i++) {
> +            if (r == 0 && i == 0) {
> +                elem->pslVirtualIsn = cpu_to_be32(ctx->irqs.offset[0]);
> +            } else {
> +                idx = ctx->irqs.offset[r] + i -
> adapter->guest->irq_base_offset;
> +                elem->applicationVirtualIsnBitmap[idx / 8] |= 0x80 >>
> (idx % 8);
> +            }
> +        }
> +    }
> +    elem->common.amr = cpu_to_be64(amr);
> +    elem->common.wed = cpu_to_be64(wed);
> +
> +    disable_afu_irqs(ctx);
> +
> +    rc = cxl_h_attach_process(ctx->afu->guest->handle, elem,
> +                &ctx->process_token, &mmio_addr, &mmio_size);
> +    if (rc == H_SUCCESS) {
> +        if (ctx->master || !ctx->afu->pp_psa) {
> +            ctx->psn_phys = ctx->afu->psn_phys;
> +            ctx->psn_size = ctx->afu->adapter->ps_size;
> +        } else {
> +            ctx->psn_phys = mmio_addr;
> +            ctx->psn_size = mmio_size;
> +        }
> +        if (ctx->afu->pp_psa && mmio_size &&
> +            ctx->afu->pp_size == 0) {
> +            /*
> +             * There's no property in the device tree to read the
> +             * pp_size. We only find out at the 1st attach.
> +             * Compared to bare-metal, it is too late and we
> +             * should really lock here. However, on powerVM,
> +             * pp_size is really only used to display in /sys.
> +             * Being discussed with pHyp for their next release.
> +             */
> +            ctx->afu->pp_size = mmio_size;
> +        }
> +        /* from PAPR: process element is bytes 4-7 of process token */
> +        ctx->external_pe = ctx->process_token & 0xFFFFFFFF;
> +        pr_devel("CXL pe=%i is known as %i for pHyp, mmio_size=%#llx",
> +            ctx->pe, ctx->external_pe, ctx->psn_size);
> +        ctx->pe_inserted = true;
> +        enable_afu_irqs(ctx);
> +    }
> +
> +    free_page((u64)elem);
> +    return rc;
> +}
> +
> +static int guest_attach_process(struct cxl_context *ctx, bool kernel,
> u64 wed, u64 amr)
> +{
> +    pr_devel("in %s\n", __func__);
> +
> +    ctx->kernel = kernel;
> +    if (ctx->afu->current_mode == CXL_MODE_DIRECTED)
> +        return attach_afu_directed(ctx, wed, amr);
> +
> +    /* dedicated mode not supported on FW840 */
> +
> +    return -EINVAL;
> +}
> +
> +static int detach_afu_directed(struct cxl_context *ctx)
> +{
> +    if (!ctx->pe_inserted)
> +        return 0;
> +    if (cxl_h_detach_process(ctx->afu->guest->handle, ctx->process_token))
> +        return -1;
> +    return 0;
> +}
> +
> +static int guest_detach_process(struct cxl_context *ctx)
> +{
> +    pr_devel("in %s\n", __func__);
> +    trace_cxl_detach(ctx);
> +
> +    if (!cxl_ops->link_ok(ctx->afu->adapter))
> +        return -EIO;
> +
> +    if (ctx->afu->current_mode == CXL_MODE_DIRECTED)
> +        return detach_afu_directed(ctx);
> +
> +    return -EINVAL;
> +}
> +
> +static void guest_release_afu(struct device *dev)
> +{
> +    struct cxl_afu *afu = to_cxl_afu(dev);
> +
> +    pr_devel("%s\n", __func__);
> +
> +    idr_destroy(&afu->contexts_idr);
> +
> +    kfree(afu->guest);
> +    kfree(afu);
> +}
> +
> +ssize_t cxl_guest_read_afu_vpd(struct cxl_afu *afu, void *buf, size_t len)
> +{
> +    return guest_collect_vpd(NULL, afu, buf, len);
> +}
> +
> +#define ERR_BUFF_MAX_COPY_SIZE PAGE_SIZE
> +static ssize_t guest_afu_read_err_buffer(struct cxl_afu *afu, char *buf,
> +                    loff_t off, size_t count)
> +{
> +    void *tbuf = NULL;
> +    int rc = 0;
> +
> +    tbuf = (void *) get_zeroed_page(GFP_KERNEL);
> +    if (!tbuf)
> +        return -ENOMEM;
> +
> +    rc = cxl_h_get_afu_err(afu->guest->handle,
> +                   off & 0x7,
> +                   virt_to_phys(tbuf),
> +                   count);
> +    if (rc)
> +        goto err;
> +
> +    if (count > ERR_BUFF_MAX_COPY_SIZE)
> +        count = ERR_BUFF_MAX_COPY_SIZE - (off & 0x7);
> +    memcpy(buf, tbuf, count);
> +err:
> +    free_page((u64)tbuf);
> +
> +    return rc;
> +}
> +
> +static int guest_afu_check_and_enable(struct cxl_afu *afu)
> +{
> +    return 0;
> +}
> +
> +static int activate_afu_directed(struct cxl_afu *afu)
> +{
> +    int rc;
> +
> +    dev_info(&afu->dev, "Activating AFU(%d) directed mode\n", afu->slice);
> +
> +    afu->current_mode = CXL_MODE_DIRECTED;
> +
> +    afu->num_procs = afu->max_procs_virtualised;
> +
> +    if ((rc = cxl_chardev_m_afu_add(afu)))
> +        return rc;
> +
> +    if ((rc = cxl_sysfs_afu_m_add(afu)))
> +        goto err;
> +
> +    if ((rc = cxl_chardev_s_afu_add(afu)))
> +        goto err1;
> +
> +    return 0;
> +err1:
> +    cxl_sysfs_afu_m_remove(afu);
> +err:
> +    cxl_chardev_afu_remove(afu);
> +    return rc;
> +}
> +
> +static int guest_afu_activate_mode(struct cxl_afu *afu, int mode)
> +{
> +    if (!mode)
> +        return 0;
> +    if (!(mode & afu->modes_supported))
> +        return -EINVAL;
> +
> +    if (mode == CXL_MODE_DIRECTED)
> +        return activate_afu_directed(afu);
> +
> +    if (mode == CXL_MODE_DEDICATED)
> +        dev_err(&afu->dev, "Dedicated mode not supported\n");
> +
> +    return -EINVAL;
> +}
> +
> +static int deactivate_afu_directed(struct cxl_afu *afu)
> +{
> +    dev_info(&afu->dev, "Deactivating AFU(%d) directed mode\n",
> afu->slice);
> +
> +    afu->current_mode = 0;
> +    afu->num_procs = 0;
> +
> +    cxl_sysfs_afu_m_remove(afu);
> +    cxl_chardev_afu_remove(afu);
> +
> +    cxl_ops->afu_reset(afu);
> +
> +    return 0;
> +}
> +
> +static int guest_afu_deactivate_mode(struct cxl_afu *afu, int mode)
> +{
> +    if (!mode)
> +        return 0;
> +    if (!(mode & afu->modes_supported))
> +        return -EINVAL;
> +
> +    if (mode == CXL_MODE_DIRECTED)
> +        return deactivate_afu_directed(afu);
> +    return 0;
> +}
> +
> +static int guest_afu_reset(struct cxl_afu *afu)
> +{
> +    pr_devel("AFU(%d) reset request\n", afu->slice);
> +    return cxl_h_reset_afu(afu->guest->handle);
> +}
> +
> +static int guest_map_slice_regs(struct cxl_afu *afu)
> +{
> +    if (!(afu->p2n_mmio = ioremap(afu->guest->p2n_phys,
> afu->guest->p2n_size))) {
> +        dev_err(&afu->dev, "Error mapping AFU(%d) MMIO regions\n",
> +            afu->slice);
> +        return -ENOMEM;
> +    }
> +    return 0;
> +}
> +
> +static void guest_unmap_slice_regs(struct cxl_afu *afu)
> +{
> +    if (afu->p2n_mmio)
> +        iounmap(afu->p2n_mmio);
> +}
> +
> +static bool guest_link_ok(struct cxl *cxl)
> +{
> +    return true;
> +}
> +
> +static int afu_properties_look_ok(struct cxl_afu *afu)
> +{
> +    if (afu->pp_irqs < 0) {
> +        dev_err(&afu->dev, "Unexpected per-process minimum interrupt
> value\n");
> +        return -EINVAL;
> +    }
> +
> +    if (afu->max_procs_virtualised < 1) {
> +        dev_err(&afu->dev, "Unexpected max number of processes
> virtualised value\n");
> +        return -EINVAL;
> +    }
> +
> +    if (afu->crs_len < 0) {
> +        dev_err(&afu->dev, "Unexpected configuration record size
> value\n");
> +        return -EINVAL;
> +    }
> +
> +    return 0;
> +}
> +
> +int cxl_guest_init_afu(struct cxl *adapter, int slice, struct
> device_node *afu_np)
> +{
> +    struct cxl_afu *afu;
> +    bool free = true;
> +    int rc;
> +
> +    pr_devel("in %s - AFU(%d)\n", __func__, slice);
> +    if (!(afu = cxl_alloc_afu(adapter, slice)))
> +        return -ENOMEM;
> +
> +    if (!(afu->guest = kzalloc(sizeof(struct cxl_afu_guest),
> GFP_KERNEL))) {
> +        kfree(afu);
> +        return -ENOMEM;
> +    }
> +
> +    if ((rc = dev_set_name(&afu->dev, "afu%i.%i",
> +                      adapter->adapter_num,
> +                      slice)))
> +        goto err1;
> +
> +    adapter->slices++;
> +
> +    if ((rc = cxl_of_read_afu_handle(afu, afu_np)))
> +        goto err1;
> +
> +    if ((rc = cxl_ops->afu_reset(afu)))
> +        goto err1;
> +
> +    if ((rc = cxl_of_read_afu_properties(afu, afu_np)))
> +        goto err1;
> +
> +    if ((rc = afu_properties_look_ok(afu)))
> +        goto err1;
> +
> +    if ((rc = guest_map_slice_regs(afu)))
> +        goto err1;
> +
> +    if ((rc = guest_register_serr_irq(afu)))
> +        goto err2;
> +
> +    /*
> +     * After we call this function we must not free the afu directly, even
> +     * if it returns an error!
> +     */
> +    if ((rc = cxl_register_afu(afu)))
> +        goto err_put1;
> +
> +    if ((rc = cxl_sysfs_afu_add(afu)))
> +        goto err_put1;
> +
> +    /*
> +     * pHyp doesn't expose the programming models supported by the
> +     * AFU. pHyp currently only supports directed mode. If it adds
> +     * dedicated mode later, this version of cxl has no way to
> +     * detect it. So we'll initialize the driver, but the first
> +     * attach will fail.
> +     * Being discussed with pHyp to do better (likely new property)
> +     */
> +    if (afu->max_procs_virtualised == 1)
> +        afu->modes_supported = CXL_MODE_DEDICATED;
> +    else
> +        afu->modes_supported = CXL_MODE_DIRECTED;
> +
> +    if ((rc = cxl_afu_select_best_mode(afu)))
> +        goto err_put2;
> +
> +    adapter->afu[afu->slice] = afu;
> +
> +    afu->enabled = true;
> +
> +    return 0;
> +
> +err_put2:
> +    cxl_sysfs_afu_remove(afu);
> +err_put1:
> +    device_unregister(&afu->dev);
> +    free = false;
> +    guest_release_serr_irq(afu);
> +err2:
> +    guest_unmap_slice_regs(afu);
> +err1:
> +    if (free) {
> +        kfree(afu->guest);
> +        kfree(afu);
> +    }
> +    return rc;
> +}
> +
> +void cxl_guest_remove_afu(struct cxl_afu *afu)
> +{
> +    pr_devel("in %s - AFU(%d)\n", __func__, afu->slice);
> +
> +    if (!afu)
> +        return;
> +
> +    cxl_sysfs_afu_remove(afu);
> +
> +    spin_lock(&afu->adapter->afu_list_lock);
> +    afu->adapter->afu[afu->slice] = NULL;
> +    spin_unlock(&afu->adapter->afu_list_lock);
> +
> +    cxl_context_detach_all(afu);
> +    cxl_ops->afu_deactivate_mode(afu, afu->current_mode);
> +    guest_release_serr_irq(afu);
> +    guest_unmap_slice_regs(afu);
> +
> +    device_unregister(&afu->dev);
> +}
> +
> +static void free_adapter(struct cxl *adapter)
> +{
> +    struct irq_avail *cur;
> +    int i;
> +
> +    if (adapter->guest->irq_avail) {
> +        for (i = 0; i < adapter->guest->irq_nranges; i++) {
> +            cur = &adapter->guest->irq_avail[i];
> +            kfree(cur->bitmap);
> +        }
> +        kfree(adapter->guest->irq_avail);
> +    }
> +    kfree(adapter->guest->status);
> +    cxl_remove_adapter_nr(adapter);
> +    kfree(adapter->guest);
> +    kfree(adapter);
> +}
> +
> +static int properties_look_ok(struct cxl *adapter)
> +{
> +    /* The absence of this property means that the operational
> +     * status is unknown or okay
> +     */
> +    if (strlen(adapter->guest->status) &&
> +        strcmp(adapter->guest->status, "okay")) {
> +        pr_err("ABORTING:Bad operational status of the device\n");
> +        return -EINVAL;
> +    }
> +
> +    return 0;
> +}
> +
> +ssize_t cxl_guest_read_adapter_vpd(struct cxl *adapter,    void *buf,
> size_t len)
> +{
> +    return guest_collect_vpd(adapter, NULL, buf, len);
> +}
> +
> +void cxl_guest_remove_adapter(struct cxl *adapter)
> +{
> +    pr_devel("in %s\n", __func__);
> +
> +    cxl_sysfs_adapter_remove(adapter);
> +
> +    device_unregister(&adapter->dev);
> +}
> +
> +static void release_adapter(struct device *dev)
> +{
> +    free_adapter(to_cxl_adapter(dev));
> +}
> +
> +struct cxl *cxl_guest_init_adapter(struct device_node *np, struct
> platform_device *pdev)
> +{
> +    struct cxl *adapter;
> +    bool free = true;
> +    int rc;
> +
> +    if (!(adapter = cxl_alloc_adapter()))
> +        return ERR_PTR(-ENOMEM);
> +
> +    if (!(adapter->guest = kzalloc(sizeof(struct cxl_guest),
> GFP_KERNEL))) {
> +        free_adapter(adapter);
> +        return ERR_PTR(-ENOMEM);
> +    }
> +
> +    adapter->slices = 0;
> +    adapter->guest->pdev = pdev;
> +    adapter->dev.parent = &pdev->dev;
> +    adapter->dev.release = release_adapter;
> +    dev_set_drvdata(&pdev->dev, adapter);
> +
> +    if ((rc = cxl_of_read_adapter_handle(adapter, np)))
> +        goto err1;
> +
> +    if ((rc = cxl_of_read_adapter_properties(adapter, np)))
> +        goto err1;
> +
> +    if ((rc = properties_look_ok(adapter)))
> +        goto err1;
> +
> +    /*
> +     * After we call this function we must not free the adapter directly,
> +     * even if it returns an error!
> +     */
> +    if ((rc = cxl_register_adapter(adapter)))
> +        goto err_put1;
> +
> +    if ((rc = cxl_sysfs_adapter_add(adapter)))
> +        goto err_put1;
> +
> +    return adapter;
> +
> +err_put1:
> +    device_unregister(&adapter->dev);
> +    free = false;
> +err1:
> +    if (free)
> +        free_adapter(adapter);
> +    return ERR_PTR(rc);
> +}
> +
> +const struct cxl_backend_ops cxl_guest_ops = {
> +    .module = THIS_MODULE,
> +    .adapter_reset = guest_reset,
> +    .alloc_one_irq = guest_alloc_one_irq,
> +    .release_one_irq = guest_release_one_irq,
> +    .alloc_irq_ranges = guest_alloc_irq_ranges,
> +    .release_irq_ranges = guest_release_irq_ranges,
> +    .setup_irq = NULL,
> +    .handle_psl_slice_error = guest_handle_psl_slice_error,
> +    .psl_interrupt = guest_psl_irq,
> +    .ack_irq = guest_ack_irq,
> +    .attach_process = guest_attach_process,
> +    .detach_process = guest_detach_process,
> +    .link_ok = guest_link_ok,
> +    .release_afu = guest_release_afu,
> +    .afu_read_err_buffer = guest_afu_read_err_buffer,
> +    .afu_check_and_enable = guest_afu_check_and_enable,
> +    .afu_activate_mode = guest_afu_activate_mode,
> +    .afu_deactivate_mode = guest_afu_deactivate_mode,
> +    .afu_reset = guest_afu_reset,
> +    .afu_cr_read8 = guest_afu_cr_read8,
> +    .afu_cr_read16 = guest_afu_cr_read16,
> +    .afu_cr_read32 = guest_afu_cr_read32,
> +    .afu_cr_read64 = guest_afu_cr_read64,
> +};
> diff --git a/drivers/misc/cxl/main.c b/drivers/misc/cxl/main.c
> index 927ba5a..b3c3ebf 100644
> --- a/drivers/misc/cxl/main.c
> +++ b/drivers/misc/cxl/main.c
> @@ -285,9 +285,6 @@ static int __init init_cxl(void)
>   {
>       int rc = 0;
>
> -    if (!cpu_has_feature(CPU_FTR_HVMODE))
> -        return -EPERM;
> -
>       if ((rc = cxl_file_init()))
>           return rc;
>
> @@ -296,8 +293,14 @@ static int __init init_cxl(void)
>       if ((rc = register_cxl_calls(&cxl_calls)))
>           goto err;
>
> -    cxl_ops = &cxl_native_ops;
> -    if ((rc = pci_register_driver(&cxl_pci_driver)))
> +    if (cpu_has_feature(CPU_FTR_HVMODE)) {
> +        cxl_ops = &cxl_native_ops;
> +        rc = pci_register_driver(&cxl_pci_driver);
> +    } else {
> +        cxl_ops = &cxl_guest_ops;
> +        rc = platform_driver_register(&cxl_of_driver);
> +    }
> +    if (rc)
>           goto err1;
>
>       return 0;
> @@ -312,7 +315,10 @@ err:
>
>   static void exit_cxl(void)
>   {
> -    pci_unregister_driver(&cxl_pci_driver);
> +    if (cpu_has_feature(CPU_FTR_HVMODE))
> +        pci_unregister_driver(&cxl_pci_driver);
> +    else
> +        platform_driver_unregister(&cxl_of_driver);
>
>       cxl_debugfs_exit();
>       cxl_file_exit();
> diff --git a/drivers/misc/cxl/of.c b/drivers/misc/cxl/of.c
> new file mode 100644
> index 0000000..edc4583
> --- /dev/null
> +++ b/drivers/misc/cxl/of.c
> @@ -0,0 +1,513 @@
> +/*
> + * Copyright 2015 IBM Corp.
> + *
> + * This program is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU General Public License
> + * as published by the Free Software Foundation; either version
> + * 2 of the License, or (at your option) any later version.
> + */
> +
> +#include <linux/kernel.h>
> +#include <linux/module.h>
> +#include <linux/platform_device.h>
> +#include <linux/slab.h>
> +#include <linux/of_address.h>
> +#include <linux/of_platform.h>
> +
> +#include "cxl.h"
> +
> +
> +static const __be32 *read_prop_string(const struct device_node *np,
> +                const char *prop_name)
> +{
> +    const __be32 *prop;
> +
> +    prop = of_get_property(np, prop_name, NULL);
> +    if (cxl_verbose && prop)
> +        pr_info("%s: %s\n", prop_name, (char *) prop);
> +    return prop;
> +}
> +
> +static const __be32 *read_prop_dword(const struct device_node *np,
> +                const char *prop_name, u32 *val)
> +{
> +    const __be32 *prop;
> +
> +    prop = of_get_property(np, prop_name, NULL);
> +    if (prop)
> +        *val = be32_to_cpu(prop[0]);
> +    if (cxl_verbose && prop)
> +        pr_info("%s: %#x (%u)\n", prop_name, *val, *val);
> +    return prop;
> +}
> +
> +static const __be64 *read_prop64_dword(const struct device_node *np,
> +                const char *prop_name, u64 *val)
> +{
> +    const __be64 *prop;
> +
> +    prop = of_get_property(np, prop_name, NULL);
> +    if (prop)
> +        *val = be64_to_cpu(prop[0]);
> +    if (cxl_verbose && prop)
> +        pr_info("%s: %#llx (%llu)\n", prop_name, *val, *val);
> +    return prop;
> +}
> +
> +
> +static int read_handle(struct device_node *np, u64 *handle)
> +{
> +    const __be32 *prop;
> +    u64 size;
> +
> +    /* Get address and size of the node */
> +    prop = of_get_address(np, 0, &size, NULL);
> +    if (size)
> +        return -EINVAL;
> +
> +    /* Helper to read a big number; size is in cells (not bytes) */
> +    *handle = of_read_number(prop, of_n_addr_cells(np));
> +    return 0;
> +}
> +
> +static int read_phys_addr(struct device_node *np, char *prop_name,
> +            struct cxl_afu *afu)
> +{
> +    int i, len, entry_size, naddr, nsize, type;
> +    u64 addr, size;
> +    const __be32 *prop;
> +
> +    naddr = of_n_addr_cells(np);
> +    nsize = of_n_size_cells(np);
> +
> +    prop = of_get_property(np, prop_name, &len);
> +    if (prop) {
> +        entry_size = naddr + nsize;
> +        for (i = 0; i < (len / 4); i += entry_size, prop += entry_size) {
> +            type = be32_to_cpu(prop[0]);
> +            addr = of_read_number(prop, naddr);
> +            size = of_read_number(&prop[naddr], nsize);
> +            switch (type) {
> +            case 0: /* unit address */
> +                afu->guest->handle = addr;
> +                break;
> +            case 1: /* p2 area */
> +                afu->guest->p2n_phys += addr;
> +                afu->guest->p2n_size = size;
> +                break;
> +            case 2: /* problem state area */
> +                afu->psn_phys += addr;
> +                afu->adapter->ps_size = size;
> +                break;
> +            default:
> +                pr_err("Invalid address type %d found in %s property of
> AFU\n",
> +                    type, prop_name);
> +                return -EINVAL;
> +            }
> +            if (cxl_verbose)
> +                pr_info("%s: %#x %#llx (size %#llx)\n",
> +                    prop_name, type, addr, size);
> +        }
> +    }
> +    return 0;
> +}
> +
> +static int read_vpd(struct cxl *adapter, struct cxl_afu *afu)
> +{
> +    char vpd[256];
> +    int rc;
> +    size_t len = sizeof(vpd);
> +
> +    memset(vpd, 0, len);
> +
> +    if (adapter)
> +        rc = cxl_guest_read_adapter_vpd(adapter, vpd, len);
> +    else
> +        rc = cxl_guest_read_afu_vpd(afu, vpd, len);
> +
> +    if (rc > 0) {
> +        cxl_dump_debug_buffer(vpd, rc);
> +        rc = 0;
> +    }
> +    return rc;
> +}
> +
> +int cxl_of_read_afu_handle(struct cxl_afu *afu, struct device_node
> *afu_np)
> +{
> +    if (read_handle(afu_np, &afu->guest->handle))
> +        return -EINVAL;
> +    pr_devel("AFU handle: 0x%.16llx\n", afu->guest->handle);
> +
> +    return 0;
> +}
> +
> +int cxl_of_read_afu_properties(struct cxl_afu *afu, struct device_node
> *np)
> +{
> +    int i, len, rc;
> +    char *p;
> +    const __be32 *prop;
> +    u16 device_id, vendor_id;
> +    u32 val = 0, class_code;
> +
> +    /* Properties are read in the same order as listed in PAPR */
> +
> +    if (cxl_verbose) {
> +        pr_info("Dump of the 'ibm,coherent-platform-function' node
> properties:\n");
> +
> +        prop = of_get_property(np, "compatible", &len);
> +        i = 0;
> +        while (i < len) {
> +            p = (char *) prop + i;
> +            pr_info("compatible: %s\n", p);
> +            i += strlen(p) + 1;
> +        }
> +        read_prop_string(np, "name");
> +    }
> +
> +    rc = read_phys_addr(np, "reg", afu);
> +    if (rc)
> +        return rc;
> +
> +    rc = read_phys_addr(np, "assigned-addresses", afu);
> +    if (rc)
> +        return rc;
> +
> +    if (afu->psn_phys == 0)
> +        afu->psa = false;
> +    else
> +        afu->psa = true;
> +
> +    if (cxl_verbose) {
> +        read_prop_string(np, "ibm,loc-code");
> +        read_prop_string(np, "device_type");
> +    }
> +
> +    read_prop_dword(np, "ibm,#processes", &afu->max_procs_virtualised);
> +
> +    if (cxl_verbose) {
> +        read_prop_dword(np, "ibm,scratchpad-size", &val);
> +        read_prop_dword(np, "ibm,programmable", &val);
> +        read_prop_string(np, "ibm,phandle");
> +        read_vpd(NULL, afu);
> +    }
> +
> +    read_prop_dword(np, "ibm,max-ints-per-process",
> &afu->guest->max_ints);
> +    afu->irqs_max = afu->guest->max_ints;
> +
> +    prop = read_prop_dword(np, "ibm,min-ints-per-process", &afu->pp_irqs);
> +    if (prop) {
> +        /* One extra interrupt for the PSL interrupt is already
> +         * included. Remove it now to keep only AFU interrupts and
> +         * match the native case.
> +         */
> +        afu->pp_irqs--;
> +    }
> +
> +    if (cxl_verbose) {
> +        read_prop_dword(np, "ibm,max-ints", &val);
> +        read_prop_dword(np, "ibm,vpd-size", &val);
> +    }
> +
> +    read_prop64_dword(np, "ibm,error-buffer-size", &afu->eb_len);
> +    afu->eb_offset = 0;
> +
> +    if (cxl_verbose)
> +        read_prop_dword(np, "ibm,config-record-type", &val);
> +
> +    read_prop64_dword(np, "ibm,config-record-size", &afu->crs_len);
> +    afu->crs_offset = 0;
> +
> +    read_prop_dword(np, "ibm,#config-records", &afu->crs_num);
> +
> +    if (cxl_verbose) {
> +        for (i = 0; i < afu->crs_num; i++) {
> +            rc = cxl_ops->afu_cr_read16(afu, i, PCI_DEVICE_ID,
> +                        &device_id);
> +            if (!rc)
> +                pr_info("record %d - device-id: %#x\n",
> +                    i, device_id);
> +            rc = cxl_ops->afu_cr_read16(afu, i, PCI_VENDOR_ID,
> +                        &vendor_id);
> +            if (!rc)
> +                pr_info("record %d - vendor-id: %#x\n",
> +                    i, vendor_id);
> +            rc = cxl_ops->afu_cr_read32(afu, i, PCI_CLASS_REVISION,
> +                        &class_code);
> +            if (!rc) {
> +                class_code >>= 8;
> +                pr_info("record %d - class-code: %#x\n",
> +                    i, class_code);
> +            }
> +        }
> +
> +        read_prop_dword(np, "ibm,function-number", &val);
> +        read_prop_dword(np, "ibm,privileged-function", &val);
> +        read_prop_dword(np, "vendor-id", &val);
> +        read_prop_dword(np, "device-id", &val);
> +        read_prop_dword(np, "revision-id", &val);
> +        read_prop_dword(np, "class-code", &val);
> +        read_prop_dword(np, "subsystem-vendor-id", &val);
> +        read_prop_dword(np, "subsystem-id", &val);
> +    }
> +    /*
> +     * if "ibm,process-mmio" doesn't exist then per-process mmio is
> +     * not supported
> +     */
> +    val = 0;
> +    prop = read_prop_dword(np, "ibm,process-mmio", &val);
> +    if (prop && val == 1)
> +        afu->pp_psa = true;
> +    else
> +        afu->pp_psa = false;
> +
> +    if (cxl_verbose) {
> +        read_prop_dword(np, "ibm,supports-aur", &val);
> +        read_prop_dword(np, "ibm,supports-csrp", &val);
> +        read_prop_dword(np, "ibm,supports-prr", &val);
> +    }
> +
> +    prop = read_prop_dword(np, "ibm,function-error-interrupt", &val);
> +    if (prop)
> +        afu->serr_hwirq = val;
> +
> +    pr_devel("AFU handle: %#llx\n", afu->guest->handle);
> +    pr_devel("p2n_phys: %#llx (size %#llx)\n",
> +        afu->guest->p2n_phys, afu->guest->p2n_size);
> +    pr_devel("psn_phys: %#llx (size %#llx)\n",
> +        afu->psn_phys, afu->adapter->ps_size);
> +    pr_devel("Max number of processes virtualised=%i\n",
> +        afu->max_procs_virtualised);
> +    pr_devel("Per-process irqs min=%i, max=%i\n", afu->pp_irqs,
> +         afu->irqs_max);
> +    pr_devel("Slice error interrupt=%#lx\n", afu->serr_hwirq);
> +
> +    return 0;
> +}
> +
> +static int read_adapter_irq_config(struct cxl *adapter, struct
> device_node *np)
> +{
> +    const __be32 *ranges;
> +    int len, nranges, i;
> +    struct irq_avail *cur;
> +
> +    ranges = of_get_property(np, "interrupt-ranges", &len);
> +    if (ranges == NULL || len < (2 * sizeof(int)))
> +        return -EINVAL;
> +
> +    /*
> +     * encoded array of two cells per entry, each cell encoded as
> +     * with encode-int
> +     */
> +    nranges = len / (2 * sizeof(int));
> +    if (nranges == 0 || (nranges * 2 * sizeof(int)) != len)
> +        return -EINVAL;
> +
> +    adapter->guest->irq_avail = kzalloc(nranges * sizeof(struct
> irq_avail),
> +                        GFP_KERNEL);
> +    if (adapter->guest->irq_avail == NULL)
> +        return -ENOMEM;
> +
> +    adapter->guest->irq_base_offset = be32_to_cpu(ranges[0]);
> +    for (i = 0; i < nranges; i++) {
> +        cur = &adapter->guest->irq_avail[i];
> +        cur->offset = be32_to_cpu(ranges[i * 2]);
> +        cur->range  = be32_to_cpu(ranges[i * 2 + 1]);
> +        cur->bitmap = kcalloc(BITS_TO_LONGS(cur->range),
> +                sizeof(*cur->bitmap), GFP_KERNEL);
> +        if (cur->bitmap == NULL)
> +            goto err;
> +        if (cur->offset < adapter->guest->irq_base_offset)
> +            adapter->guest->irq_base_offset = cur->offset;
> +        if (cxl_verbose)
> +            pr_info("available IRQ range: %#lx-%#lx (%lu)\n",
> +                cur->offset, cur->offset + cur->range - 1,
> +                cur->range);
> +    }
> +    adapter->guest->irq_nranges = nranges;
> +    spin_lock_init(&adapter->guest->irq_alloc_lock);
> +
> +    return 0;
> +err:
> +    for (i--; i >= 0; i--) {
> +        cur = &adapter->guest->irq_avail[i];
> +        kfree(cur->bitmap);
> +    }
> +    kfree(adapter->guest->irq_avail);
> +    adapter->guest->irq_avail = NULL;
> +    return -ENOMEM;
> +}
> +
> +int cxl_of_read_adapter_handle(struct cxl *adapter, struct device_node
> *np)
> +{
> +    if (read_handle(np, &adapter->guest->handle))
> +        return -EINVAL;
> +    pr_devel("Adapter handle: 0x%.16llx\n", adapter->guest->handle);
> +
> +    return 0;
> +}
> +
> +int cxl_of_read_adapter_properties(struct cxl *adapter, struct
> device_node *np)
> +{
> +    int rc, len, naddr, i;
> +    char *p;
> +    const __be32 *prop;
> +    u32 val = 0;
> +
> +    /* Properties are read in the same order as listed in PAPR */
> +
> +    naddr = of_n_addr_cells(np);
> +
> +    if (cxl_verbose) {
> +        pr_info("Dump of the 'ibm,coherent-platform-facility' node
> properties:\n");
> +
> +        read_prop_dword(np, "#address-cells", &val);
> +        read_prop_dword(np, "#size-cells", &val);
> +
> +        prop = of_get_property(np, "compatible", &len);
> +        i = 0;
> +        while (i < len) {
> +            p = (char *) prop + i;
> +            pr_info("compatible: %s\n", p);
> +            i += strlen(p) + 1;
> +        }
> +        read_prop_string(np, "name");
> +        read_prop_string(np, "model");
> +
> +        prop = of_get_property(np, "reg", NULL);
> +        if (prop) {
> +            pr_info("reg: addr:%#llx size:%#x\n",
> +                of_read_number(prop, naddr),
> +                be32_to_cpu(prop[naddr]));
> +        }
> +
> +        read_prop_string(np, "ibm,loc-code");
> +    }
> +
> +    if ((rc = read_adapter_irq_config(adapter, np)))
> +        return rc;
> +
> +    if (cxl_verbose) {
> +        read_prop_string(np, "device_type");
> +        read_prop_string(np, "ibm,phandle");
> +    }
> +
> +    prop = read_prop_dword(np, "ibm,caia-version", &val);
> +    if (prop) {
> +        adapter->caia_major = (val & 0xFF00) >> 8;
> +        adapter->caia_minor = val & 0xFF;
> +    }
> +
> +    prop = read_prop_dword(np, "ibm,psl-revision", &val);
> +    if (prop)
> +        adapter->psl_rev = val;
> +
> +    prop = read_prop_string(np, "status");
> +    if (prop) {
> +        adapter->guest->status = kasprintf(GFP_KERNEL, "%s", (char *)
> prop);
> +        if (adapter->guest->status == NULL)
> +            return -ENOMEM;
> +    }
> +
> +    prop = read_prop_dword(np, "vendor-id", &val);
> +    if (prop)
> +        adapter->guest->vendor = val;
> +
> +    prop = read_prop_dword(np, "device-id", &val);
> +    if (prop)
> +        adapter->guest->device = val;
> +
> +    if (cxl_verbose) {
> +        read_prop_dword(np, "ibm,privileged-facility", &val);
> +        read_prop_dword(np, "revision-id", &val);
> +        read_prop_dword(np, "class-code", &val);
> +    }
> +
> +    prop = read_prop_dword(np, "subsystem-vendor-id", &val);
> +    if (prop)
> +        adapter->guest->subsystem_vendor = val;
> +
> +    prop = read_prop_dword(np, "subsystem-id", &val);
> +    if (prop)
> +        adapter->guest->subsystem = val;
> +
> +    if (cxl_verbose)
> +        read_vpd(adapter, NULL);
> +
> +    return 0;
> +}
> +
> +static int cxl_of_remove(struct platform_device *pdev)
> +{
> +    struct cxl *adapter;
> +    int afu;
> +
> +    adapter = dev_get_drvdata(&pdev->dev);
> +    for (afu = 0; afu < adapter->slices; afu++)
> +        cxl_guest_remove_afu(adapter->afu[afu]);
> +
> +    cxl_guest_remove_adapter(adapter);
> +    return 0;
> +}
> +
> +static void cxl_of_shutdown(struct platform_device *pdev)
> +{
> +    cxl_of_remove(pdev);
> +}
> +
> +int cxl_of_probe(struct platform_device *pdev)
> +{
> +    struct device_node *np = NULL;
> +    struct device_node *afu_np = NULL;
> +    struct cxl *adapter = NULL;
> +    int ret;
> +    int slice, slice_ok;
> +
> +    pr_devel("in %s\n", __func__);
> +
> +    np = pdev->dev.of_node;
> +    if (np == NULL)
> +        return -ENODEV;
> +
> +    /* init adapter */
> +    adapter = cxl_guest_init_adapter(np, pdev);
> +    if (IS_ERR(adapter)) {
> +        dev_err(&pdev->dev, "guest_init_adapter failed: %li\n",
> PTR_ERR(adapter));
> +        return PTR_ERR(adapter);
> +    }
> +
> +    /* init afu */
> +    slice_ok = 0;
> +    for (afu_np = NULL, slice = 0; (afu_np = of_get_next_child(np,
> afu_np)); slice++) {
> +        if ((ret = cxl_guest_init_afu(adapter, slice, afu_np)))
> +            dev_err(&pdev->dev, "AFU %i failed to initialise: %i\n",
> +                slice, ret);
> +        else
> +            slice_ok++;
> +    }
> +
> +    if (slice_ok == 0) {
> +        dev_info(&pdev->dev, "No active AFU");
> +        adapter->slices = 0;
> +    }
> +
> +    if (afu_np)
> +        of_node_put(afu_np);
> +    return 0;
> +}
> +
> +static const struct of_device_id cxl_of_match[] = {
> +    { .compatible = "ibm,coherent-platform-facility",},
> +    {},
> +};
> +MODULE_DEVICE_TABLE(of, cxl_of_match);
> +
> +struct platform_driver cxl_of_driver = {
> +    .driver = {
> +        .name = "cxl_of",
> +        .of_match_table = cxl_of_match,
> +        .owner = THIS_MODULE
> +    },
> +    .probe = cxl_of_probe,
> +    .remove = cxl_of_remove,
> +    .shutdown = cxl_of_shutdown,
> +};



More information about the Linuxppc-dev mailing list