Hi All,

I am also in favor of CUDA suffixes. CUDA containing toolchain is only needed when you need CUDA aware MPI (It does not work with Intel as far as I know). In other words, if you want to use GPUs in different boxes.

Sincerely,

Balazs


On 19/03/2018 15:43, Joachim Hein wrote:
Hi,

I am currently installing tensorflow via easybuild (I assume many of us do these days) and am trying to understand EasyBuild’s ideas on toolchains supporting cuda.

I looked at TensorFlow-1.5.0-goolfc-2017b-Python-3.6.3.eb, which builds ontop of a toolchain containing GCC, Cuda (installed as a compiler module), and OpenMPI, Blas, FFTW etc.

I now noticed that there is a new TensorFlow-1.6.0-foss-2018a-Python-3.6.4-CUDA-9.1.85.eb, which is accepted into the development branch (PR 6016).   This builds ontop a “vanilla” foss-2018a toolchain, using a  Cuda and cuDNN modules installed as a core module (system compiler).

I am wondering how do we want to organise us in future?  Do we want to continue with the goolfc idea or do we go for a “core” cuda and cuDNN?  I feel this needs standardising soonish.  It is also something I feel I need to document for my users, who want to build their own cuda based software.  What models should be loaded to build software.

Any comments, how we take this further?

Best wishes
   Joachim


--
HPC consultant
HPC/VSC Support and System Administration
Computing Center
ULB/VUB
Avenue Adolphe Buyllaan 91 - CP 197
1050 Brussels
Belgium

Reply via email to