[PATCH kernel v2 6/6] powerpc/powernv/ioda: Allocate indirect TCE levels on demand

Alexey Kardashevskiy aik at ozlabs.ru
Thu Jun 21 17:16:46 AEST 2018


On Thu, 21 Jun 2018 12:03:21 +1000
David Gibson <david at gibson.dropbear.id.au> wrote:

> On Sun, Jun 17, 2018 at 09:14:28PM +1000, Alexey Kardashevskiy wrote:
> > At the moment we allocate the entire TCE table, twice (hardware part and
> > userspace translation cache). This normally works as we normally have
> > contigous memory and the guest will map entire RAM for 64bit DMA.
> > 
> > However if we have sparse RAM (one example is a memory device), then
> > we will allocate TCEs which will never be used as the guest only maps
> > actual memory for DMA. If it is a single level TCE table, there is nothing
> > we can really do but if it a multilevel table, we can skip allocating
> > TCEs we know we won't need.
> > 
> > This adds ability to allocate only first level, saving memory.
> > 
> > This changes iommu_table::free() to avoid allocating of an extra level;
> > iommu_table::set() will do this when needed.
> > 
> > This adds @alloc parameter to iommu_table::exchange() to tell the callback
> > if it can allocate an extra level; the flag is set to "false" for
> > the realmode KVM handlers of H_PUT_TCE hcalls and the callback returns
> > H_TOO_HARD.
> > 
> > This still requires the entire table to be counted in mm::locked_vm.
> > 
> > To be conservative, this only does on-demand allocation when
> > the usespace cache table is requested which is the case of VFIO.
> > 
> > The example math for a system replicating a powernv setup with NVLink2
> > in a guest:
> > 16GB RAM mapped at 0x0
> > 128GB GPU RAM window (16GB of actual RAM) mapped at 0x244000000000
> > 
> > the table to cover that all with 64K pages takes:
> > (((0x244000000000 + 0x2000000000) >> 16)*8)>>20 = 4556MB
> > 
> > If we allocate only necessary TCE levels, we will only need:
> > (((0x400000000 + 0x400000000) >> 16)*8)>>20 = 4MB (plus some for indirect
> > levels).
> > 
> > Signed-off-by: Alexey Kardashevskiy <aik at ozlabs.ru>
> > ---
> > Changes:
> > v2:
> > * fixed bug in cleanup path which forced the entire table to be
> > allocated right before destroying
> > * added memory allocation error handling pnv_tce()
> > ---
> >  arch/powerpc/include/asm/iommu.h              |  7 ++-
> >  arch/powerpc/platforms/powernv/pci.h          |  6 ++-
> >  arch/powerpc/kvm/book3s_64_vio_hv.c           |  4 +-
> >  arch/powerpc/platforms/powernv/pci-ioda-tce.c | 69 ++++++++++++++++++++-------
> >  arch/powerpc/platforms/powernv/pci-ioda.c     |  8 ++--
> >  drivers/vfio/vfio_iommu_spapr_tce.c           |  2 +-
> >  6 files changed, 69 insertions(+), 27 deletions(-)
> > 
> > diff --git a/arch/powerpc/include/asm/iommu.h b/arch/powerpc/include/asm/iommu.h
> > index 4bdcf22..daa3ee5 100644
> > --- a/arch/powerpc/include/asm/iommu.h
> > +++ b/arch/powerpc/include/asm/iommu.h
> > @@ -70,7 +70,7 @@ struct iommu_table_ops {
> >  			unsigned long *hpa,
> >  			enum dma_data_direction *direction);
> >  
> > -	__be64 *(*useraddrptr)(struct iommu_table *tbl, long index);
> > +	__be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc);
> >  #endif
> >  	void (*clear)(struct iommu_table *tbl,
> >  			long index, long npages);
> > @@ -122,10 +122,13 @@ struct iommu_table {
> >  	__be64 *it_userspace; /* userspace view of the table */
> >  	struct iommu_table_ops *it_ops;
> >  	struct kref    it_kref;
> > +	int it_nid;
> >  };
> >  
> > +#define IOMMU_TABLE_USERSPACE_ENTRY_RM(tbl, entry) \
> > +		((tbl)->it_ops->useraddrptr((tbl), (entry), false))
> >  #define IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry) \
> > -		((tbl)->it_ops->useraddrptr((tbl), (entry)))
> > +		((tbl)->it_ops->useraddrptr((tbl), (entry), true))
> >  
> >  /* Pure 2^n version of get_order */
> >  static inline __attribute_const__
> > diff --git a/arch/powerpc/platforms/powernv/pci.h b/arch/powerpc/platforms/powernv/pci.h
> > index 5e02408..1fa5590 100644
> > --- a/arch/powerpc/platforms/powernv/pci.h
> > +++ b/arch/powerpc/platforms/powernv/pci.h
> > @@ -267,8 +267,10 @@ extern int pnv_tce_build(struct iommu_table *tbl, long index, long npages,
> >  		unsigned long attrs);
> >  extern void pnv_tce_free(struct iommu_table *tbl, long index, long npages);
> >  extern int pnv_tce_xchg(struct iommu_table *tbl, long index,
> > -		unsigned long *hpa, enum dma_data_direction *direction);
> > -extern __be64 *pnv_tce_useraddrptr(struct iommu_table *tbl, long index);
> > +		unsigned long *hpa, enum dma_data_direction *direction,
> > +		bool alloc);
> > +extern __be64 *pnv_tce_useraddrptr(struct iommu_table *tbl, long index,
> > +		bool alloc);
> >  extern unsigned long pnv_tce_get(struct iommu_table *tbl, long index);
> >  
> >  extern long pnv_pci_ioda2_table_alloc_pages(int nid, __u64 bus_offset,
> > diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c
> > index db0490c..05b4865 100644
> > --- a/arch/powerpc/kvm/book3s_64_vio_hv.c
> > +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c
> > @@ -200,7 +200,7 @@ static long kvmppc_rm_tce_iommu_mapped_dec(struct kvm *kvm,
> >  {
> >  	struct mm_iommu_table_group_mem_t *mem = NULL;
> >  	const unsigned long pgsize = 1ULL << tbl->it_page_shift;
> > -	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> > +	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RM(tbl, entry);
> >  
> >  	if (!pua)
> >  		/* it_userspace allocation might be delayed */
> > @@ -264,7 +264,7 @@ static long kvmppc_rm_tce_iommu_do_map(struct kvm *kvm, struct iommu_table *tbl,
> >  {
> >  	long ret;
> >  	unsigned long hpa = 0;
> > -	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> > +	__be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RM(tbl, entry);
> >  	struct mm_iommu_table_group_mem_t *mem;
> >  
> >  	if (!pua)
> > diff --git a/arch/powerpc/platforms/powernv/pci-ioda-tce.c b/arch/powerpc/platforms/powernv/pci-ioda-tce.c
> > index 36c2eb0..a7debfb 100644
> > --- a/arch/powerpc/platforms/powernv/pci-ioda-tce.c
> > +++ b/arch/powerpc/platforms/powernv/pci-ioda-tce.c
> > @@ -48,7 +48,7 @@ static __be64 *pnv_alloc_tce_level(int nid, unsigned int shift)
> >  	return addr;
> >  }
> >  
> > -static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx)
> > +static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc)
> >  {
> >  	__be64 *tmp = user ? tbl->it_userspace : (__be64 *) tbl->it_base;
> >  	int  level = tbl->it_indirect_levels;
> > @@ -57,7 +57,20 @@ static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx)
> >  
> >  	while (level) {
> >  		int n = (idx & mask) >> (level * shift);
> > -		unsigned long tce = be64_to_cpu(tmp[n]);
> > +		unsigned long tce;
> > +
> > +		if (tmp[n] == 0) {
> > +			__be64 *tmp2;
> > +
> > +			if (!alloc)
> > +				return NULL;
> > +
> > +			tmp2 = pnv_alloc_tce_level(tbl->it_nid,
> > +					ilog2(tbl->it_level_size) + 3);  
> 
> Can this allocation fail?  If it does you'll crash as you dereference
> NULL just below.


Oh. I fixed both comments and I lost the fix in rebase, I'll repost :(



--
Alexey
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 833 bytes
Desc: OpenPGP digital signature
URL: <http://lists.ozlabs.org/pipermail/linuxppc-dev/attachments/20180621/94ef3406/attachment.sig>


More information about the Linuxppc-dev mailing list