AlexanderSerov opened a new issue #19521:
URL: https://github.com/apache/incubator-mxnet/issues/19521


   ## Problem statement
   Our team using mxnet for training and for inference. In recent time we have 
intention to run inference on Android devices so we compile mxnet using android 
ndk and it works fine. Now we have intention to accelerate inference on mobile 
devices using [android NN 
Api](https://developer.android.com/ndk/guides/neuralnetworks) which android 
support since version 8.1. This Api serve as common interface to hardware 
GPU/Accelerator drivers and provide api in the form of operators ( 
ANEURALNETWORKS_CONV_2D, ANEURALNETWORKS_AVERAGE_POOL_2D...).
   
   ## Proposed solutions
   My task is to implement a proxy between mxnet and android nn using subgraph 
api and actually i already on half a way. I already implement selector, 
subgraph property, register opearator, and impement addition of major operator 
to android nn model based on partitioned graph. The design is similar to 
TensorRT subgraph but we don't use onnx as interim. So the question is, is it 
wise to implement subgraph for running inference on mobile device using 
framework which initially not have intention to run inference on mobile. I mean 
mxnet size in apk is about 150 MB that is pretty thick. I use mxnet 1.7. Will 
there a lightweight version of mxnet in future like TFlite for tensorflow? 
Also, any suggestions and thoughts about more appropriate solution for our 
problem are welcome!


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to