Hi Zhen, On 17/04/15 16:57, Zhen Zhang wrote: > And as a (maybe) off-topic discussion, take this solution as an > example, how is the communication between processes done? Will the > data from one card be move t the main memory and moved towards > another cards? Or the cards can go the PCIe Bus directly? > > P.S. The direct communication seems to be called the GPUDirect > technology.
There is a fork of PyFR that can exploit CUDA-aware-MPI. This permits PyFR to pass CUDA device pointers directly to MPI functions and have the MPI library handle the copying. This may, or may not, use peer-to-peer GPU copies or perhaps even GPUDirect over RDMA. If there is any performance benefit depends heavily on the problem being solved, the platform it is being solved on, and the MPI library. As PyFR already tries very hard to overlap communication with computation, and because for large transfer sizes most MPI libraries will fall back to copying via the host, the benefit is usually quite small. Regards, Freddie. -- You received this message because you are subscribed to the Google Groups "PyFR Mailing List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send an email to [email protected]. Visit this group at http://groups.google.com/group/pyfrmailinglist. For more options, visit https://groups.google.com/d/optout.
signature.asc
Description: OpenPGP digital signature
