KellenSunderland commented on a change in pull request #11325: Added TensorRT runtime integration URL: https://github.com/apache/incubator-mxnet/pull/11325#discussion_r202182189
########## File path: src/executor/exec_pass.h ########## @@ -198,6 +198,27 @@ Graph InferStorageType(Graph&& graph, StorageTypeVector&& storage_type_inputs = StorageTypeVector(), const std::string& storage_type_attr_key = ""); +/*! \brief The default storage type inference function, which assigns all undefined + * storage types to kDefaultStorage. If all of input and output storage types + * are kDefaultStorage, DispatchMode::kFCompute is assigned to dispatch_mode. Otherwise, + * DispatchMode::kFComputeFallback is assigned to dispatch_mode. + */ +bool DefaultStorageType(const nnvm::NodeAttrs& attrs, + const int dev_mask, + DispatchMode* dispatch_mode, + std::vector<int> *iattr, + std::vector<int> *oattr); + +/*! + * \brief Replace subgraphs by TRT (forward only) + */ +Graph ReplaceSubgraph(Graph&& g, + const std::unordered_set<nnvm::Node*>& set_subgraph, + std::unordered_map<std::string, NDArray>* const params_map); + +std::vector<std::unordered_set<nnvm::Node*>> GetTrtCompatibleSubsets(const Graph& g, Review comment: This seems TRT specific. Should we include this method only in TRT builds (i.e. should we wrap in an #if MXNET_USE_TENSORRT) ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services