On Fri, Sep 8, 2017 at 4:45 PM, Gilles Gouaillardet
wrote:
> can you
> ./mpicc -showme -I/usr/local/cuda-8.0.61/lib64 -lcuda test_cuda_aware.c -o
> myapp
> and double check -lcuda is *after* -lopen-pal ?
gcc -I/usr/local/cuda-8.0.61/lib64 -lcuda test_cuda_aware.c
To solve the undefined references to cudaMalloc and cudaFree, you need
to link the CUDA runtime. So you should replace -lcuda by -lcudart.
For the OPENCL undefined references, I don't know where those are coming
from ... could it be that hwloc is compiling OpenCL support but not
adding
On 09/08/2017 8:16 AM, Marco Atzeri wrote:
please reply in the mailing list
Oops! My apologies. I'm not used to a mailing list without the reply-to
set to the mailing list.
Can a version of open mpi be built using x86_64-w64-mingw32-gcc so
that it will work with code compiled with
In my experience, POSIX is much more reliable than Sys5. Sys5 depends on
the value of shmmax, which is often set to a small fraction of node
memory. I've probably seen the error described on
http://verahill.blogspot.com/2012/04/solution-to-nwchem-shmmax-too-small.html
with NWChem a 1000 times
Joseph,
Thanks for sharing this !
sysv is imho the worst option because if something goes really wrong, Open MPI
might leave some shared memory segments behind when a job crashes. From that
perspective, leaving a big file in /tmp can be seen as the lesser evil.
That being said, there might be
We are currently discussing internally how to proceed with this issue on
our machine. We did a little survey to see the setup of some of the
machines we have access to, which includes an IBM, a Bull machine, and
two Cray XC40 machines. To summarize our findings:
1) On the Cray systems, both
On 08/09/2017 14:15, Llelan D. wrote:
Thank you for your reply.
please reply in the mailing list
The x86_64-w64-mingw32-gcc set are compilers written to compile code in
the Cygwin64 environment that do not need the Cygwin64 dll to run (just
like the MinGW environment). Please be more
Tom --
If you're going to upgrade, can you upgrade to the latest Open MPI (2.1.1)?
I.e., unless you have a reason for wanting to stay back at an already-old
version, you might as well upgrade to the latest latest latest to give you the
longest shelf life.
I mention this because we are
What you probably want to do is add --cpu-list a,b,c... to each mpirun command,
where each one lists the cores you want to assign to that job.
> On Sep 8, 2017, at 6:46 AM, twu...@goodyear.com wrote:
>
>
> I posted this question last year and we ended up not upgrading to the newer
> openmpi.
It isn’t an issue as there is nothing wrong with OMPI. Your method of joining
the allocation is a problem. What you have done is to create a job step that
has only 1 slot/node. We have no choice but to honor that constraint and run
within it.
What you should be doing is to use salloc to create
I posted this question last year and we ended up not upgrading to the newer
openmpi. Now I need to change to openmpi 1.10.5 and have the same issue.
Specifically, using 1.4.2, I can run two 12 core jobs on a 24 core node and the
processes would bind to cores and only have 1 process per core.
Hi,
I have a successful installation of Nvidia drivers and cuda which is
confirmed by
"nvcc -V" and "nvidia-smi".
after configuring the openmpi with
"./configure --with-cuda --prefix=/home/umashankar/softwares/openmpi-2.0.3"
"make all install"
and after exporting paths ,
I ended up with an
can you
./mpicc -showme -I/usr/local/cuda-8.0.61/lib64 -lcuda test_cuda_aware.c -o myapp
and double check -lcuda is *after* -lopen-pal ?
Cheers,
Gilles
On Fri, Sep 8, 2017 at 7:40 PM, Nilesh Kokane wrote:
> On Fri, Sep 8, 2017 at 4:08 PM, Nilesh Kokane
On 08/09/2017 02:38, Llelan D. wrote:
Windows 10 64bit, Cygwin64, openmpi 1.10.7-1 (dev, c, c++, fortran),
x86_64-w64-mingw32-gcc 6.3.0-1 (core, gcc, g++, fortran)
I am compiling the standard "hello_c.c" example with *mgicc* configured
to use the Cygwin installed MinGW gcc compiler:
$
On Fri, Sep 8, 2017 at 4:08 PM, Nilesh Kokane wrote:
> On Fri, Sep 8, 2017 at 3:33 PM, Gilles Gouaillardet
> wrote:
>>
>> Nilesh,
>>
>> Can you
>> configure --without-nvidia ...
>> And see if it helps ?
>
> No, I need Nvidia cuda support.
On Fri, Sep 8, 2017 at 3:33 PM, Gilles Gouaillardet
wrote:
>
> Nilesh,
>
> Can you
> configure --without-nvidia ...
> And see if it helps ?
No, I need Nvidia cuda support.
//Nilesh Kokane
___
users mailing list
Hi,
I have a system running openmpi programs over archlinux. I had the programs
compiled and running on July when I was using the version 1.10.4 or .7 of
openmpi if I remember correctly. Just recently I updated the openmpi
version to 2.1.1 and tried running a compiled program and ran correctly.
Nilesh,
Can you
configure --without-nvidia ...
And see if it helps ?
Cheers,
Gilles
Nilesh Kokane wrote:
>Hello,
>
>
>How can I compile openmpi without the support of open-cl?
>
>
>The only link I could find is [1], but openmpi doesn't configure this option.
>
>The
For the time being, you can
srun --ntasks-per-node 24 --jobid=...
When joining the allocation.
This use case looks a bit convoluted to me, so i am not even sure we should
consider there is a bug in Open MPI.
Ralph, any thoughts ?
Cheers,
Gilles
Gilles Gouaillardet wrote:
Hello,
How can I compile openmpi without the support of open-cl?
The only link I could find is [1], but openmpi doesn't configure this
option.
The reason why I'm trying to build openmpi without open-cl is it throws the
following errors even with the nvidia installed opencl.
./mpicc
Thanks, now i can reproduce the issue
Cheers,
Gilles
On 9/8/2017 5:20 PM, Maksym Planeta wrote:
I run start an interactive allocation and I just noticed that the problem
happens, when I join this allocation from another shell.
Here is how I join:
srun --pty --x11 --jobid=$(squeue -u
by any chance, would you be able to test the latest openmpi 2.1.2rc3 ?
OpenMPI 2.1.0 is the latest on our cluster.
--
Regards,
Maksym Planeta
smime.p7s
Description: S/MIME Cryptographic Signature
___
users mailing list
users@lists.open-mpi.org
I run start an interactive allocation and I just noticed that the problem
happens, when I join this allocation from another shell.
Here is how I join:
srun --pty --x11 --jobid=$(squeue -u $USER -o %A | tail -n 1) bash
And here is how I create the allocation:
srun --pty --nodes 8
Maxsym,
can you please post your sbatch script ?
fwiw, i am unable to reproduce the issue with the latest v2.x from github.
by any chance, would you be able to test the latest openmpi 2.1.2rc3 ?
Cheers,
Gilles
On 9/8/2017 4:19 PM, Maksym Planeta wrote:
Indeed mpirun shows slots=1 per
Indeed mpirun shows slots=1 per node, but I create allocation with
--ntasks-per-node 24, so I do have all cores of the node allocated.
When I use srun I can get all the cores.
On 09/07/2017 02:12 PM, r...@open-mpi.org wrote:
> My best guess is that SLURM has only allocated 2 slots, and we
25 matches
Mail list logo