bgawrych commented on issue #20066:
URL: 
https://github.com/apache/incubator-mxnet/issues/20066#issuecomment-809986931


   
https://github.com/apache/incubator-mxnet/blob/5722f8b38af58c5a296e46ca695bfaf7cff85040/src/operator/numpy/np_bincount_op.cu#L103-L111
   
   
https://github.com/apache/incubator-mxnet/blob/5722f8b38af58c5a296e46ca695bfaf7cff85040/src/operator/numpy/np_bincount_op.cu#L146-L152
   
   Template mechanism here create Kernel with int16 data type, because it can 
be returned by MSHADOW_TYPE_SWITCH - as cudnn doesn't support int16 you can 
define new macro (or search for existing one) which doesn't support int16 e.g ( 
MSHADOW_TYPE_SWITCH_NO_INT16) and replace it (I believe it's not the only one). 
However to avoid potential bugs for GPU implementations should be check whether 
input/output is int16 datatype and if it is, there should be fallback mechanism 
implemented to CPU implementation. Something like:
   
https://github.com/apache/incubator-mxnet/blob/66a65924f03e6e62ca0619afb02e2a674fcccbfd/src/operator/tensor/elemwise_sum.cc#L117-L121
  


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to