Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Maxime Boissonneault
Hi Jeff, Le 2014-08-15 17:50, Jeff Squyres (jsquyres) a écrit : On Aug 15, 2014, at 5:39 PM, Maxime Boissonneault wrote: Correct. Can it be because torque (pbs_mom) is not running on the head node and mpiexec attempts to contact it ? Not for Open

Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Jeff Squyres (jsquyres)
On Aug 15, 2014, at 5:39 PM, Maxime Boissonneault wrote: > Correct. > > Can it be because torque (pbs_mom) is not running on the head node and > mpiexec attempts to contact it ? Not for Open MPI's mpiexec, no. Open MPI's mpiexec (mpirun -- they're the

Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Maxime Boissonneault
Correct. Can it be because torque (pbs_mom) is not running on the head node and mpiexec attempts to contact it ? Maxime Le 2014-08-15 17:31, Joshua Ladd a écrit : But OMPI 1.8.x does run the ring_c program successfully on your compute node, right? The error only happens on the front-end

Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Joshua Ladd
But OMPI 1.8.x does run the ring_c program successfully on your compute node, right? The error only happens on the front-end login node if I understood you correctly. Josh On Fri, Aug 15, 2014 at 5:20 PM, Maxime Boissonneault < maxime.boissonnea...@calculquebec.ca> wrote: > Here are the

Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Maxime Boissonneault
Here are the requested files. In the archive, you will find the output of configure, make, make install as well as the config.log, the environment when running ring_c and the ompi_info --all. Just for a reminder, the ring_c example compiled and ran, but produced no output when running and

Re: [OMPI users] Running a hybrid MPI+openMP program

2014-08-15 Thread Reuti
Hi, Am 15.08.2014 um 19:56 schrieb Oscar Mojica: > Yes, my installation of Open MPI is SGE-aware. I got the following > > [oscar@compute-1-2 ~]$ ompi_info | grep grid > MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.6.2) Fine. > I'm a bit slow and I didn't understand

Re: [OMPI users] Running a hybrid MPI+openMP program

2014-08-15 Thread Oscar Mojica
Hi ReutiYes, my installation of Open MPI is SGE-aware. I got the following[oscar@compute-1-2 ~]$ ompi_info | grep grid MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.6.2)I'm a bit slow and I didn't understand the las part of your message. So i made a test trying to solve

Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Maxime Boissonneault
Hi, I solved the warning that appeared with OpenMPI 1.6.5 on the login node. I increased the registrable memory. Now, with OpenMPI 1.6.5, it does not give any warning. Yet, with OpenMPI 1.8.1 and OpenMPI 1.8.2rc4, it still exits with error code 65 and does not produce the normal output. I

Re: [OMPI users] Segmentation fault in OpenMPI 1.8.1

2014-08-15 Thread Maxime Boissonneault
Hi Josh, The ring_c example does not work on our login node : [mboisson@helios-login1 examples]$ mpiexec -np 10 ring_c [mboisson@helios-login1 examples]$ echo $? 65 [mboisson@helios-login1 examples]$ echo $LD_LIBRARY_PATH

Re: [hwloc-users] [WARNING: A/V UNSCANNABLE] hwloc error

2014-08-15 Thread Andrej Prsa
> Most people don't care about cache when binding with MPI, so you may > just ignore the issue and hide the message by setting > HWLOC_HIDE_ERRORS=1 in the environment. It may work fine (assuming > MPIs don't have troubles with asymmetric topologies where there are > some missing L3). We do see

Re: [hwloc-users] [WARNING: A/V UNSCANNABLE] hwloc error

2014-08-15 Thread Andrej Prsa
Hi Brice, > Your kernel looks recent enough, can you try upgrading your BIOS ? You > have version 3.0b and there's a 3.5 version at > http://www.supermicro.com/aplus/motherboard/opteron6000/sr56x0/h8qg6-f.cfm Flashing bios is not the easiest option for us since I'd need to bring down the whole

Re: [hwloc-users] [WARNING: A/V UNSCANNABLE] hwloc error

2014-08-15 Thread Brice Goglin
Hello, Your platform reports buggy L3 cache locality information. This is very common on AMD 62xx and 63xx platforms unfortunately. You have 8 L3 caches (one per 6-core NUMA node, two per socket), but the platform report 11 L3 caches instead: Socket s1, 2 and 4 report one L3 above 2 cores, one

[hwloc-users] hwloc error

2014-08-15 Thread Andrej Prsa
Hi guys, * hwloc has encountered what looks like an error from the operating system. * * L3 (cpuset 0x03f0) intersects with NUMANode (P#0 cpuset 0x003f) without inclusion! * Error occurred in topology.c line 940