tracking of PCI address space

Kumar Gala galak at kernel.crashing.org
Thu Apr 9 07:53:52 EST 2009


On Apr 8, 2009, at 4:49 PM, Ira Snyder wrote:

> On Wed, Apr 08, 2009 at 03:53:55PM -0500, Kumar Gala wrote:
>> I was wondering if we have anything that tracks regions associated  
>> with
>> the "inbound" side of a pci_bus.
>>
>> What I mean is on embedded PPC we have window/mapping registers for  
>> both
>> inbound (accessing memory on the SoC) and outbound (access PCI device
>> MMIO, IO etc).  The combination of the inbound & outbound convey what
>> exists in the PCI address space vs CPU physical address space (and  
>> how to
>> map from one to the other).  Today in the PPC land we only attach
>> outbound windows to the pci_bus.  So technically the inbound side
>> information (like what subset of physical memory is visible on the  
>> PCI
>> bus) seems to be lost.
>>
>
> To the best of my knowledge there is no API to set inbound windows in
> Linux. I've been implementing a virtio-over-PCI driver which needs the
> inbound windows. I set them up myself during driver probe, using
> get_immrbase() to get the IMMR registers. This board is a PCI Slave /
> Agent, it doesn't even have PCI support compiled into the kernel.

I'm not concerned explicitly about setting up inbound windows, its  
more about have a consistent view of the PCI address space which may  
be different from the CPU physical address space.

I'm working on code to actually setup the inbound windows on 85xx/86xx  
class devices (based on dma-ranges property in the device tree).  As I  
was thinking about this I realized that the send of ranges/dma-ranges  
in the .dts and what we map to outbound vs inbound changes if we an  
agent or host.

- k



More information about the Linuxppc-dev mailing list