How to support 3GB pci address?

maillist.kernel maillist.kernel at gmail.com
Sat Dec 13 15:56:26 EST 2008


Thanks for all the suggestions and Comments!
In the system,  the total memory size is less than 4GB.   I want to know how to map the 3GB pci address to the kernel ,
and how can my driver access all the pci device. Usually we have only 1GB kernel address, minus the 896MB
for memory map, we can only use the 128M for ioremap. I can adjust the PAGE_OFFSET to a lower value, but it's not enough.

>One can mmap() a PCI BAR from userspace, in which case the mapping comes
>out of the "max userspace size" pool instead of the "all ioremap()s" pool. 
>The userspace pool is per processes.  So while having four kernel drivers
>each call ioremap(..., 1GB) will never work, it is possible to have four
>userspace processes each call mmap("/sys/bus/pci.../resource", 1GB) and
>have it work.

There are many pci devices in the system and every pci device has only several tens of MB, so how can I call the mmap("/sys/bus/pci.../resource", 1GB) 
and how can I use it by my drivers ?

Thanks again for all your help!




 
On Fri, 12 Dec 2008, Kumar Gala wrote:
> On Dec 12, 2008, at 3:04 AM, Trent Piepho wrote:
> > On Thu, 11 Dec 2008, Kumar Gala wrote:
> >  > On Dec 11, 2008, at 10:07 PM, Trent Piepho wrote:
> >  >  > On Thu, 11 Dec 2008, Kumar Gala wrote:
> >  >  >  > The 36-bit support is current (in tree) in complete.  Work is in 
> >  >  >  > add swiotlb support to PPC which will generically enable what you 
> >  >  > 
> >  >  > Don't the ATMU windows in the pcie controller serve as a IOMMU, making
> >  >  > swiotlb
> >  >  > unnecessary and wasteful?
> >  > 
> >  > Nope.  You have no way to tell when to switch a window as you have no 
> >  > idea
> >  > when a device might DMA data.
> > 
> > Isn't that what dma_alloc_coherent() and dma_map_single() are for?
>
> Nope.  How would manipulate the PCI ATMU?

Umm, out_be32()?  Why would it be any different than other iommu
implementations, like the pseries one for example?

Just define set a of fsl dma ops that use an inbound ATMU window if they
need to.  The only issue would be if you have a 32-bit device with multiple
concurrent DMA buffers scattered over  > 32 bits of address space and run
out of ATMU windows.  But other iommu implementations have that same
limitation.  You just have to try harder to allocate GFP_DMA memory that
doesn't need an ATMU window or create larger contiguous bounce buffer to
replace scattered smaller buffers.

> > It sounded like the original poster was talking about having 3GB of PCI
> > BARs.  How does swiotlb even enter the picture for that?
>
> It wasn't clear how much system memory they wanted.  If they can fit their 
> entire memory map for PCI addresses in 4G of address space (this includes all 
> of system DRAM) than they don't need anything special.

Why the need to fit the entire PCI memory map into the lower 4G?  What
issue is there with mapping a PCI BAR above 4G if you have 36-bit support?

Putting system memory below 4GB is only an issue if you're talking about
DMA.  For mapping a PCI BAR, what does it doesn't matter?

The problem I see with having large PCI BARs, is that the max userspace
process size plus low memory plus all ioremap()s must be less than 4GB.  If
one wants to call ioremap(..., 3GB), then only 1 GB is left for userspace
plus low memory.  That's not very much.

One can mmap() a PCI BAR from userspace, in which case the mapping comes
out of the "max userspace size" pool instead of the "all ioremap()s" pool. 
The userspace pool is per processes.  So while having four kernel drivers
each call ioremap(..., 1GB) will never work, it is possible to have four
userspace processes each call mmap("/sys/bus/pci.../resource", 1GB) and
have it work.

> >  >From what I've read about swiotlb, it is a hack that allows one to do DMA
> > with 32-bit PCI devices on 64-bit systems that lack an IOMMU.  It reserves
> > a large block of RAM under 32-bits (technically it uses GFP_DMA) and doles
> > this out to drivers that allocate DMA memory.
>
> correct.  It bounce buffers the DMAs to a 32-bit dma'ble region and copies 
> to/from the  >32-bit address.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ozlabs.org/pipermail/linuxppc-dev/attachments/20081213/f82b4458/attachment.htm>


More information about the Linuxppc-dev mailing list