zixuanweeei opened a new issue #17898: LSTM and GRU layers without DNNL enabled 
give wrong gradients
URL: https://github.com/apache/incubator-mxnet/issues/17898
 
 
   ## Description
   Currently, we have two implementations of RNN layers on the CPU backend, 
which are
   
   + Native fusion implementation,
   + The fusion enabled by DNNL library 
(https://intel.github.io/mkl-dnn/dev_guide_rnn.html).
   
   Both of them can be invoked from `mx.sym.RNN`, `mx.rnn.FusedRNNCell`, 
`mx.gluon.rnn.LSTM/GRU/RNN`. The fusion of DNNL provides more efficient Forward 
and Backward, while the native one gives a backup for some devices or 
environments that cannot use DNNL library.
   
   Recently, we have found that there are some problems leading to the wrong 
gradients' calculation of the native implementation. Just tracking the issue 
here, and it will be fixed ASAP.
   
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

Reply via email to