Re: [OMPI users] mpirun error

2014-11-04 Thread jfsanchez
On Tuesday, November 04, 2014 11:06:06 PM Jeff Squyres wrote: > mpirun -np 4 ./test Oh man, THANKS for your quick answer. LOL, nice, now it's running. III Escuela Internacional de Invierno en la UCI del 17 al 28 de febrero del 2014. Ver www.uci.cu

Re: [OMPI users] mpirun error

2014-11-04 Thread Jeff Squyres (jsquyres)
On Nov 4, 2014, at 5:56 PM, jfsanchez wrote: > mpirun -np 4 test Try: mpirun -np 4 ./test To specifically get the "test" executable in your directory (vs. /bin/test, which OMPI may have found in your PATH). -- Jeff Squyres jsquy...@cisco.com For corporate

[OMPI users] mpirun error

2014-11-04 Thread jfsanchez
Hello, I'm very new using OpenMPI, so I was testing it and wrote this program: //main.c #include #include int main(int argc, char **argv) { int rank; char hostname[256]; MPI_Init(,); MPI_Comm_rank(MPI_COMM_WORLD, ); gethostname(hostname,255);

Re: [OMPI users] What could cause a segfault in OpenMPI?

2014-11-04 Thread Jeff Squyres (jsquyres)
Looks like it's failing in the openib BTL setup. Can you send the info listed here? http://www.open-mpi.org/community/help/ On Nov 4, 2014, at 1:10 PM, Saliya Ekanayake wrote: > Hi, > > I am using OpenMPI 1.8.1 in a Linux cluster that we recently setup. It builds >

Re: [OMPI users] change in behaviour 1.6 -> 1.8 under sge

2014-11-04 Thread Ralph Castain
I confirmed that things are working as intended. If you have 12 cores on a machine, and you do mpirun -map-by socket:PE=2 we will execute 6 copies of foo on the node because 12 cores/2pe/core = 6 procs. As I said, we believe the prior series were doing this incorrectly, and the patch used on

Re: [OMPI users] What could cause a segfault in OpenMPI?

2014-11-04 Thread Saliya Ekanayake
Hi Howard, I just tried with 1.8.3. as well and it produces the same error. We have another cluster where both versions work fine, which is why I was curious as what kind of things could cause this. Thank you, Saliya On Tue, Nov 4, 2014 at 1:31 PM, Howard Pritchard wrote:

Re: [OMPI users] What could cause a segfault in OpenMPI?

2014-11-04 Thread Howard Pritchard
Hello Saliya, Would you mind trying to reproduce the problem using the latest 1.8 release - 1.8.3? Thanks, Howard 2014-11-04 11:10 GMT-07:00 Saliya Ekanayake : > Hi, > > I am using OpenMPI 1.8.1 in a Linux cluster that we recently setup. It > builds fine, but when I try to

[OMPI users] What could cause a segfault in OpenMPI?

2014-11-04 Thread Saliya Ekanayake
Hi, I am using OpenMPI 1.8.1 in a Linux cluster that we recently setup. It builds fine, but when I try to run even the simplest hello.c program it'll cause a segfault. Any suggestions on how to correct this? The steps I did and error message are below. 1. Built OpenMPI 1.8.1 on the cluster. The

[OMPI users] ipath_userinit errors

2014-11-04 Thread Michael Di Domenico
I'm getting the below message on my cluster(s). It seems to only happen when I try to use more then 64 nodes (16-cores each). The clusters are running RHEL 6.5 with Slurm and Openmpi-1.6.5 with PSM. I'm using the OFED versions included with RHEL for infiniband support. ipath_userinit:

Re: [OMPI users] change in behaviour 1.6 -> 1.8 under sge

2014-11-04 Thread Dave Love
Ralph Castain writes: > If you only have one allocated PE on a node, then mpirun will > correctly tell you that it can’t launch with PE>1 as there aren’t > enough resources to meet your request. IIRC, we may have been ignoring > this under SGE and running as many procs as

Re: [OMPI users] change in behaviour 1.6 -> 1.8 under sge

2014-11-04 Thread Dave Love
I wrote: > #$ -l exclusive > export OMP_NUM_THREADS=2 > exec mpirun --loadbalance --cpus-per-proc $OMP_NUM_THREADS --np > $(($NSLOTS/$OMP_NUM_THREADS)) ... I should have said core binding is the default here [so Intel MPI