Good feedback.

On 2018/10/30 20:50, Leif Lindholm wrote:
On Tue, Oct 30, 2018 at 09:39:24AM -0300, Ard Biesheuvel wrote:
(add back the list)

Oi! Go back on holiday!

On 30 October 2018 at 09:07, Cohen, Eugene <[email protected]> wrote:
Has this patch been tested on a system that does not have coherent DMA?

It's not clear that this change would actually be faster on a system of that
type since using common buffers imply access to uncached memory.  Depending
on the access patterns the uncached memory access could be more time
consuming than cache maintenance operations.

The change/idea was based on the statement below.
  ///
/// Provides both read and write access to system memory by both the processor and a /// bus master. The buffer is coherent from both the processor's and the bus master's point of view.
  ///
  EfiPciIoOperationBusMasterCommonBuffer,

Thanks for raising case about uncached memory access. But after checking the code, for Intel VTd case https://github.com/tianocore/edk2/blob/master/IntelSiliconPkg/Feature/VTd/IntelVTdDxe/BmDma.c#L460 (or no IOMMU case https://github.com/tianocore/edk2/blob/master/MdeModulePkg/Bus/Pci/PciHostBridgeDxe/PciRootBridgeIo.c#L1567), the common buffer is just normal memory buffer. If someone can help do some test/collect some data on a system using common buffers imply access to uncached memory, that will be great.


I haven't had time to look at these patches yet.

I agree with Eugene's concern: the directional DMA routines are much
more performant on implementations with non-coherent DMA, and so
common buffers should be avoided unless we are dealing with data
structures that are truly shared between the CPU and the device.

Since this is obviously not the case here, could we please have some
numbers about the performance improvement we are talking about here?
Would it be possible to improve the IOMMU handling code instead?

We collected the data below on a platform with release image and Intel VTd enabled.

The image size of EhciDxe or XhciDxe can reduce about 120+ bytes.

EHCI without the patch:
==[ Cumulative ]========
(Times in microsec.)     Cumulative   Average     Shortest    Longest
   Name         Count     Duration    Duration    Duration    Duration
-------------------------------------------------------------------------------
S0000B00D1DF0        446        2150           4           2         963

EHCI with the patch:
==[ Cumulative ]========
(Times in microsec.)     Cumulative   Average     Shortest    Longest
   Name         Count     Duration    Duration    Duration    Duration
-------------------------------------------------------------------------------
S0000B00D1DF0        270         742           2           2          41

XHCI without the patch:
==[ Cumulative ]========
(Times in microsec.)     Cumulative   Average     Shortest    Longest
   Name         Count     Duration    Duration    Duration    Duration
-------------------------------------------------------------------------------
S0000B00D14F0        215         603           2           2          52

XHCI with the patch:
==[ Cumulative ]========
(Times in microsec.)     Cumulative   Average     Shortest    Longest
   Name         Count     Duration    Duration    Duration    Duration
-------------------------------------------------------------------------------
S0000B00D14F0         95         294           3           2          52

I believe the performance data really depends on
1. How many AsyncInterruptTransfer handlers (the number of USB keyboard and/or USB bluetooth keyboard?) 2. Data size (for flushing data from PCI controller specific address to mapped system memory address *in original code*) 3. The performance of IoMmu->SetAttribute (for example, the SetAttribute operation on Intel VTd engine caused by the unmap and map for flushing data *in original code*, the SetAttribute operation on IntelVTd engine will involve FlushPageTableMemory, InvalidatePageEntry and etc)


On an unrelated note to the concerns above:
Why has a fundamental change to the behaviour of one of the industry
standard drivers been pushed at the very end of the stable cycle?

We thought it was a simple improvement but not fundamental change before Eugene and Ard raised the concern.


Thanks,
Star


Regards,

Leif


_______________________________________________
edk2-devel mailing list
[email protected]
https://lists.01.org/mailman/listinfo/edk2-devel

Reply via email to