[ 
https://issues.apache.org/jira/browse/MXNET-11?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16395518#comment-16395518
 ] 

Chris Olivier commented on MXNET-11:
------------------------------------

I suppose the multiple threads would call the C API.  Python is a bit tricky to 
get it to do multithreading, so I wouldn't be concerned with a python entry 
point at this point.

So above, you launched many processes, loaded a model and ran inference in 
parallel?

Or you loaded a model in many threads and ran inference through those?

The main problem right now is that there's not a good way to *share* parameters 
between graphs, so that would probably be some large bulk of the work. There's 
actually several other use-cases for this, including Keras integration.

 

 

> Multithreaded Inference
> -----------------------
>
>                 Key: MXNET-11
>                 URL: https://issues.apache.org/jira/browse/MXNET-11
>             Project: Apache MXNet
>          Issue Type: Epic
>          Components: MXNet Engine
>            Reporter: Chris Olivier
>            Priority: Major
>              Labels: inference
>
> Add the ability to do multithreaded inference without using fork() or using 
> multiple copies of a given model



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@mxnet.apache.org
For additional commands, e-mail: issues-h...@mxnet.apache.org

Reply via email to