[PATCH] powerpc/npu-dma.c: Fix deadlock in mmio_invalidate
Mark Hairgrove
mhairgrove at nvidia.com
Fri Feb 16 14:11:19 AEDT 2018
On Wed, 14 Feb 2018, Alistair Popple wrote:
> > > +struct mmio_atsd_reg {
> > > + struct npu *npu;
> > > + int reg;
> > > +};
> > > +
> >
> > Is it just easier to move reg to inside of struct npu?
>
> I don't think so, struct npu is global to all npu contexts where as this is
> specific to the given invalidation. We don't have enough registers to assign
> each NPU context it's own dedicated register so I'm not sure it makes sense to
> put it there either.
>
> > > +static void acquire_atsd_reg(struct npu_context *npu_context,
> > > + struct mmio_atsd_reg mmio_atsd_reg[NV_MAX_NPUS])
> > > +{
> > > + int i, j;
> > > + struct npu *npu;
> > > + struct pci_dev *npdev;
> > > + struct pnv_phb *nphb;
> > >
> > > - /*
> > > - * The GPU requires two flush ATSDs to ensure all entries have
> > > - * been flushed. We use PID 0 as it will never be used for a
> > > - * process on the GPU.
> > > - */
> > > - if (flush)
> > > - mmio_invalidate_pid(npu, 0, true);
> > > + for (i = 0; i <= max_npu2_index; i++) {
> > > + mmio_atsd_reg[i].reg = -1;
> > > + for (j = 0; j < NV_MAX_LINKS; j++) {
> >
> > Is it safe to assume that npu_context->npdev will not change in this
> > loop? I guess it would need to be stronger than just this loop.
>
> It is not safe to assume that npu_context->npdev won't change during this loop,
> however I don't think it is a problem if it does as we only read each element
> once during the invalidation.
Shouldn't that be enforced with READ_ONCE() then?
I assume that npdev->bus can't change until after the last
pnv_npu2_destroy_context() is called for an npu. In that case, the
mmu_notifier_unregister() in pnv_npu2_release_context() will block until
mmio_invalidate() is done using npdev. That seems safe enough, but a
comment somewhere about that would be useful.
>
> There are two possibilities for how this could change. pnv_npu2_init_context()
> will add a nvlink to the npdev which will result in the TLB invalidation being
> sent to that GPU as well which should not be a problem.
>
> pnv_npu2_destroy_context() will remove the the nvlink from npdev. If it happens
> prior to this loop it should not be a problem (as the destruction will have
> already invalidated the GPU TLB). If it happens after this loop it shouldn't be
> a problem either (it will just result in an extra TLB invalidate being sent to
> this GPU).
>
> > > + npdev = npu_context->npdev[i][j];
> > > + if (!npdev)
> > > + continue;
> > > +
> > > + nphb = pci_bus_to_host(npdev->bus)->private_data;
> > > + npu = &nphb->npu;
> > > + mmio_atsd_reg[i].npu = npu;
> > > + mmio_atsd_reg[i].reg = get_mmio_atsd_reg(npu);
> > > + while (mmio_atsd_reg[i].reg < 0) {
> > > + mmio_atsd_reg[i].reg = get_mmio_atsd_reg(npu);
> > > + cpu_relax();
> >
> > A cond_resched() as well if we have too many tries?
>
> I don't think we can as the invalidate_range() function is called under the ptl
> spin-lock and is not allowed to sleep (at least according to
> include/linux/mmu_notifier.h).
>
> - Alistair
>
> > Balbir
> >
>
>
>
More information about the Linuxppc-dev
mailing list