KellenSunderland edited a comment on issue #12931: How do define onnx path for 
tensorrt integration?
URL: 
https://github.com/apache/incubator-mxnet/issues/12931#issuecomment-439750388
 
 
   @vandanavk  #13310 might take some iterating to get right, but this is also 
addressed in https://github.com/apache/incubator-mxnet/pull/12469 which is 
ready to be merged if someone can have a look.
   
   @nchafni:  First of all at the moment ONNX's proto client needs to be 
generated before compiling.  Make sure you're doing a 
   
   ```bash
    # Build ONNX
   pushd .
   echo "Installing ONNX."
   cd 3rdparty/onnx-tensorrt/third_party/onnx
   rm -rf build
   mkdir -p build
   cd build
   cmake \
           -DCMAKE_CXX_FLAGS=-I/usr/include/python${PYVER}\
           -DBUILD_SHARED_LIBS=ON ..\
           -G Ninja
   ninja -j 1 -v onnx/onnx.proto
   ninja -j 1 -v
   export LIBRARY_PATH=`pwd`:`pwd`/onnx/:$LIBRARY_PATH
   export CPLUS_INCLUDE_PATH=`pwd`:$CPLUS_INCLUDE_PATH
   popd
   
   # Build ONNX-TensorRT
   pushd .
   cd 3rdparty/onnx-tensorrt/
   mkdir -p build
   cd build
   cmake ..
   make -j$(nproc)
   export LIBRARY_PATH=`pwd`:$LIBRARY_PATH
   popd
   ```
   before compilation.
   
   I'f you've already done that and the onnx.pb.h file is present you can try 
and add it to your cpp include path so the compiler can pick it up. You could 
try building using cmake (just add -DUSE_TENSORRT).
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

Reply via email to