On 06.10.21 11:32, Andreas Messerschmid wrote:
> Hi all,
> 
> did anyone already implement an ivshmem-net link between two Linux cells
> using the ZynqMP PCIe hardware instead of a virtual PCI interface? What
> about MSI/MSI-X in this case?
> 
> Any hints/success stories on this?
> 

I've looked into that topics for a couple of ARM boards, again and
again, but so far only the (historic) AMD Seattle was fulfilling all
conditions to allow PCI device partitioning more or less easily (but
that thing had no IOMMU IIRC, this this was incomplete). On other HW,
you have a combination of these issues (or even the full list):

 - missing differentiation of PCI devices on the SMMU in front of the
   host controller
 - missing way to inject ivshmem interrupts at the point where the OS
   would expect them from a real device
 - complex PCI host controller, deviating from the generic one,
   requiring extra logic to intercept config space accesses or even more
 - things I forgot

Therefore, it is generally easier to add a virtual PCI host controller,
even if the SOC already has a real one.

Jan

-- 
Siemens AG, T RDA IOT
Corporate Competence Center Embedded Linux

-- 
You received this message because you are subscribed to the Google Groups 
"Jailhouse" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/jailhouse-dev/45a5e58c-83d9-9461-3b2d-08eda7eefa24%40siemens.com.

Reply via email to