bgawrych commented on issue #20066: URL: https://github.com/apache/incubator-mxnet/issues/20066#issuecomment-809986931
https://github.com/apache/incubator-mxnet/blob/5722f8b38af58c5a296e46ca695bfaf7cff85040/src/operator/numpy/np_bincount_op.cu#L103-L111 https://github.com/apache/incubator-mxnet/blob/5722f8b38af58c5a296e46ca695bfaf7cff85040/src/operator/numpy/np_bincount_op.cu#L146-L152 Template mechanism here create Kernel with int16 data type, because it can be returned by MSHADOW_TYPE_SWITCH - as cudnn doesn't support int16 you can define new macro (or search for existing one) which doesn't support int16 e.g ( MSHADOW_TYPE_SWITCH_NO_INT16) and replace it (I believe it's not the only one). However to avoid potential bugs for GPU implementations should be check whether input/output is int16 datatype and if it is, there should be fallback mechanism implemented to CPU implementation. Something like: https://github.com/apache/incubator-mxnet/blob/66a65924f03e6e62ca0619afb02e2a674fcccbfd/src/operator/tensor/elemwise_sum.cc#L117-L121 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
