On May 19, 2009, at 8:04 AM, Kumar Gala wrote:


On May 18, 2009, at 4:49 PM, Benjamin Herrenschmidt wrote:

On Mon, 2009-05-18 at 08:25 -0500, Kumar Gala wrote:

Part of this is how the generic swiotlb code works and part of it
was
our desire not to bloat dev archdata by adding such info that as you
say is either bus specific or conveyed in the dma addr mask.

Right but perf sucks :-)

Maybe an option is to clamp the DMA mask when it's set by the driver to
limit it to the available inbound window ?

Clamping the DMA mask is even worse than the additional indirection for us. We have valid scenarios in which we'd have 512M of outbound PCI address space and 4G of mem and thus 3.5G of inbound PCI address space. With the DMA mask we'd be limited to 2G and bouncing from 2..3.5G when we don't need to.



I think our options are to change archdata as follows:

Option 1 - just add a new data member to dev_archdata

struct dev_archdata {
       /* Optional pointer to an OF device node */
       struct device_node      *of_node;

       /* DMA operations on that device */
       struct dma_mapping_ops  *dma_ops;
       void                     *dma_data;
        dma_addr_t              direct_dma_addr;
};

Option 2 - introduce a proper container for how we use dma_data. This may just be moving the indirection from an indirection function call to an indirection data reference:

struct dma_data {
       dma_addr_t      offset;
       dma_addr_t      direct_dma_addr;
       struct          iommu_table *iommu_table;
};

struct dev_archdata {
       /* Optional pointer to an OF device node */
       struct device_node      *of_node;

       /* DMA operations on that device */
       struct dma_mapping_ops  *dma_ops;
       struct dma_data         *dma_data;
};

Personally, I prefer this. However, I understand Ben has some objection here. At a minimum, I *do* need to change dma_data to be able to contain a 64-bit number on a 32-bit platform as part of the optimization for 64-bit PCI devices. It should probably be a dma_addr_t for that, looking at how it's being used. However, that is a bit of a philosophical break from the type of dma_data being generic void *. I'm open to suggestions here.



Option 3 - use dma_data to keep the addr at which we need to bounce vs not for SWIOTLB - this has potential issues w/conflicting with dma_data being used as the dma_offset. (need to think on that a bit more). Additionally this has the benefit in that we need dma_data to be a 64-bit quantity on ppc32 w/>32-bit phys addr.

struct dev_archdata {
       /* Optional pointer to an OF device node */
       struct device_node      *of_node;

       /* DMA operations on that device */
       struct dma_mapping_ops  *dma_ops;
       dma_addr_t               dma_data;
};


This won't work. I need the dma offset that's currently stored there for 64-bit PCI devices.

It sounds like we want to do some combination of the above:

1) add a max_dma_addr field to archdata to indicate the boundary beyond which a device must bounce buffer (we currently don't have anything that requires a start/size type paradigm here - does anyone know of any cases that I don't know about here?)

2) change the type of dma_data so it can hold 64 bits on a 32-bit platform.

-Becky

_______________________________________________
Linuxppc-dev mailing list
Linuxppc-dev@ozlabs.org
https://ozlabs.org/mailman/listinfo/linuxppc-dev

Reply via email to