On 10/1/2025 1:57 PM, Jerin Jacob wrote:
> 
> 
>> -----Original Message-----
>> From: Vamsi Krishna Attunuru <[email protected]>
>> Sent: Wednesday, September 24, 2025 9:45 AM
>> To: fengchengwen <[email protected]>; [email protected];
>> [email protected]
>> Cc: [email protected]; [email protected];
>> [email protected]; [email protected];
>> [email protected]; Jerin Jacob <[email protected]>
>> Subject: RE: [EXTERNAL] Re: [RFC] lib/dma: introduce inter-process and 
>> inter-OS
>> DMA
>>
>> Hi Feng, Anatoly,
>>
>> Gentle ping for below review.
>>
>>> -----Original Message-----
>>> From: Vamsi Krishna Attunuru
>>> Sent: Monday, September 22, 2025 5:19 PM
>>> To: fengchengwen <[email protected]>; [email protected];
>>> [email protected]
>>> Cc: [email protected]; [email protected];
>>> [email protected]; [email protected];
>>> [email protected]; Jerin Jacob <[email protected]>
>>> Subject: RE: [EXTERNAL] Re: [RFC] lib/dma: introduce inter-process and
>>> inter- OS DMA
>>>
>>> Hi Feng,
>>>
>>>> Hi Vamsi, This commit change is more than discussed, it add control
>>>> API which for group management. 1. Control API: I check this commit
>>>> and Intel commit [1], it seem has a quite difference. I hope Intel
>>>> guys can express views
>>>> ZjQcmQRYFpfptBannerEnd
>>>> Hi Vamsi,
>>>>
>>>> This commit change is more than discussed, it add control API which
>>>> for group management.
>>>>
>>>> 1. Control API: I check this commit and Intel commit [1], it seem has
>>>> a quite difference.
>>>>   I hope Intel guys can express views. I prefer not add this part if
>>>> no
>>> response.
>>>
>>> This new feature needs to be securely managed through control APIs. It
>>> would be extremely helpful if the folks at Intel and you as well could
>>> provide support or inputs on this.
> 
> 
> Beyond adding Intel folks to this thread, I don't see any further steps we 
> can take to drive review at this stage.
> 
> That said, the table-based concept used in the current API may not be 
> portable, and we may need improvements here.
> 
> Based on my understanding, DMA devices used for inter process copy can be 
> classified into three categories:
> 
> Class A: Requires a pair of DMA devices (one on each end of process/domain) 
> for data transfer. Marvell DMA devices fall into this category.
> Class B: Requires only a single DMA device (one process/domain has a DMA 
> device, the other process does not). Intel DMA devices fall here.
> Class C: Other types of devices that we are not yet aware of.
> 
> Abstracting all of these under a single API will be challenging. Linux and 
> other OSes do not provide control-plane APIs for this,
> so DPDK must provide control plane mechanisms to support Class A, Class B, 
> and Class C devices.
> 
> Proposal: Split development into separate sets:
> -----------------------------------------------
> Set A: Focus only on the datapath. Assume uint16_t *src_handle and uint16_t 
> *dst_handle come from elsewhere (Class C).
> Set B: Introduce capabilities for Class A devices with portable APIs 
> (proposal below, without table concept).
> Set C: Introduce capabilities for Class B devices and relevant APIs, to be 
> added when needed.
> 
> We can merge Set A in the current release and move Set B to a next release 
> _if_ review or support for Class A devices requires more time.
> 
> @fengchengwen Thoughts? 

okay

> 
> Class A API Proposal:
> ---------------------
> These APIs are based on a new capability flag for inter-process or inter-OS 
> DMA transfers for Class A devices.
> 
> 
> /** Creates an access group for pair-type inter-process or inter-OS DMA 
> transfers. */
> int rte_dma_access_pair_group_create(const struct rte_dma_dev *dev,
>                                      rte_uuid_t process_id,
>                                      rte_uuid_t token,
>                                      uint16_t *group_id);

how about rte_dma_access_group_create(), and add pair-type as one parameter, 
and also rename process_id as domain_id

> 
> /** Destroys an access group once all participating devices have exited. */
> int rte_dma_access_pair_group_destroy(const struct rte_dma_dev *dev,
>                                       uint16_t group_id);

rte_dma_access_group_destroy()

> 
> /** Allows a device to join an existing access group using a device handle 
> and token. */
> int rte_dma_access_pair_group_join(const struct rte_dma_dev *dev,
>                                    uint16_t group_id,
>                                    rte_uuid_t process_id,
>                                    rte_uuid_t token,
>                                    rte_dma_access_pair_leave_cb_t leave_cb);

rte_dma_access_group_join()

> 
> /** Removes a device from an access group. */
> int rte_dma_access_pair_group_leave(const struct rte_dma_dev *dev,
>                                     uint16_t group_id);
> 
> /** Retrieves the source and destination handles for a given device within 
> the group. */
> int rte_dma_access_pair_gorup_src_dst_handles_get(const struct rte_dma_dev 
> *dev,
>                                             uint16_t group_id,
>                                             rte_uuid_t src_process_id,
>                                             rte_uuid_t dst_process_id,
>                                             uint16_t *src_handle,
>                                             uint16_t *dst_handle);

rte_dma_access_group_handle_get(const struct  rte_dma_dev *dev,
                                uint16_t group_id,
                                rte_uuid_t domain_id,
                                uint16_t *handle);

so user could invoke multiple time if they want to get differ domain_id's 
handle.

> 
> 
> Parameters that need explanation:
> --------------------------------
> process_id: Unique ID for the process, generated via rte_uuid_ APIs
> token: Provided by an administrative actor to grant access, similar to VFIO 
> VF token creation used in VFIO PF driver.
> leave_cb: Callback to notify when another side leaves the group
> 
> 
> Example Workflow for Class A Inter-Domain DMA Transfer:
> -------------------------------------------------------
> 
> This example demonstrates how three processes — p0, p1, and p2 — coordinate 
> inter-domain DMA transfers using pair-type(Class A) DMA devices.
> 
> Step 1: Group Creation (p0)
> Process p0 calls rte_dma_access_pair_group_create() with a unique process 
> handle and token. A group_id is returned.

I prefer the group_id is int type so it could hold such like file descriptor.

> 
> Step 2: Group Sharing
> group_id and token are shared with p1 and p2 via IPC or shared memory.
> 
> Step 3: Group Joining (p1 & p2)
> Processes p1 and p2 call rte_dma_access_pair_group_join() with their process 
> id  and the shared token from admin
> 
> Step 4: Handle Discovery
> Each process uses rte_dma_access_pair_gorup_src_dst_handles_get() to retrieve 
> source and destination handles for other processes.
> 
> Step 5: Transfer Coordination
> Using the handles, each process configures a virtual channel (vchan) and 
> initiates DMA transfers.
> 
> Step 6: Group Teardown
> When a process no longer needs to participate, it calls 
> rte_dma_access_pair_group_leave(). Other processes are notified via the 
> registered callback with rte_dma_access_pair_group_join().
> Once all devices have exited, p0 calls rte_dma_access_pair_group_destroy() to 
> clean up.
> 
> 
> For Class B: We can add new capability flag and have new set of APIs 
> rte_dma_access_master_ or so. When such devices comes or when intel wants to 
> add it
> 
> 
> 
> 
> 

Reply via email to