cuDNN 7.2 has simplified the usage of Tensor Cores for convolutions and RNNs, 
and explicit casts to `float16` can be avoided: 
https://devblogs.nvidia.com/tensor-ops-made-easier-in-cudnn/

Are there plans to expose this in MXNet already? If not, lets discuss a design 
in this issue.

@szha


[ Full content available at: 
https://github.com/apache/incubator-mxnet/issues/12463 ]
This message was relayed via gitbox.apache.org for [email protected]

Reply via email to