apeforest commented on issue #16735: Use single-bit for mask in dropout operator
URL: https://github.com/apache/incubator-mxnet/pull/16735#issuecomment-589927989
 
 
   Hi @TaoLv I reverted my last commit "
   It makes the code too bristle and also involves very complicate logic to 
check memory allocation at runtime. Here are the main reasons:
   
   (1) MKL dropout support is currently not complete. It does not work if the 
input data type is smaller than int32 and it does not support broadcast option 
(when the option axes is specified). This limitation enforces a check at 
runtime which is not possible in the InferShape function
   
   e.g. In this function, I will need to check if the dtype is greater than 
int32 in order to use a different shape for MKL Dropout. 
   
https://github.com/apache/incubator-mxnet/pull/16735/files#diff-74c4dc433970c5df31a5e2c4b57c8d71R127
   
   (2) Having different Dropout engine at runtime (based on data type and ) may 
cause inconsistency in the mixed precision case. Introducing another difference 
in mask memory allocation complicates this even further.
   
   I think we should focus on enhancing MKL Dropout so that it (1) supports all 
the different cases as non MKL dropout (2) supports bit-mask.
   
   Please let me know what you think. Thanks!
   
   Lin 
   
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

Reply via email to