beew edited a comment on issue #18879:
URL: 
https://github.com/apache/incubator-mxnet/issues/18879#issuecomment-670987830


   @szha 
   Hi, 
   
   1.without source /opt/intel/bin/compilervars.sh intel64, build with openblas 
instead of mkl (since can't find mkl root and headers) 
   python gpu examples works (from here 
https://mxnet.apache.org/versions/1.6/get_started/validate_mxnet.html)
   
   2 With source /opt/intel/bin/compilervars.sh intel6 and build with mkl then 
gpu examples hang as reported here, regardless whether JEMALLOC is turned on or 
off.  If it is turned off, got an additional warning "Significant performance 
increases can be achieved by installing and enabling gperftools or jemalloc 
development packages"
   
   So it seems that mkl somehow doesn't play well with cuda...
   
   I haven't tried the nightly branch as it seems that the prebuild mxnet 
python wheels only support cuda > 10.0 after version 1.5.1,  so instead I tried 
pip install mxnet-cu100mkl==1.5.1  I was able to create the mxnet array in gpu 
context with no problem . 
   
   P.S except for the prebuild mxnet wheels, all other tests were done with 
mxnet 1.7 from source "git checkout -b v1.7 origin/v1.7.x"
   
   


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to