bgawrych edited a comment on issue #20066:
URL: 
https://github.com/apache/incubator-mxnet/issues/20066#issuecomment-809986931


   
https://github.com/apache/incubator-mxnet/blob/5722f8b38af58c5a296e46ca695bfaf7cff85040/src/operator/numpy/np_bincount_op.cu#L103-L111
   
   
https://github.com/apache/incubator-mxnet/blob/5722f8b38af58c5a296e46ca695bfaf7cff85040/src/operator/numpy/np_bincount_op.cu#L146-L152
   
   Template mechanism here create Kernel with int16 data type, because it can 
be returned by MSHADOW_TYPE_SWITCH - as cudnn doesn't support int16 you can 
define new macro (or search for existing one) which doesn't support int16 e.g ( 
MSHADOW_TYPE_SWITCH_NO_INT16) and replace it (I believe this problem will 
occurs in other places as well). However to avoid potential bugs for GPU 
implementations should be check whether input/output is int16 datatype and if 
it is, there should be fallback mechanism implemented to CPU implementation. 
Something like:
   
https://github.com/apache/incubator-mxnet/blob/66a65924f03e6e62ca0619afb02e2a674fcccbfd/src/operator/tensor/elemwise_sum.cc#L117-L121
  


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to