Re: [OMPI users] MPI_Send doesn't work if the data >= 2GB

2010-12-07 Thread 孟宪军
Hi Bowen, Thanks very much. I had checked my writev system call, I thought it was him that caused all these bad things :) Best Regards Xianjun Meng 2010/12/8 Bowen Zhou > On 12/05/2010 10:13 PM, > >> hi, >> >> I met a question recently when I tested the MPI_send and

Re: [OMPI users] glut display 'occasionally' opens

2010-12-07 Thread Ralph Castain
FWIW: I just tested the -x option on a multi-node system and had no problem getting the value of DISPLAY to propagate. I was able to define it on the cmd line, saw it set correctly on every process, etc. This was with our devel trunk - not sure what version you are using. On Dec 7, 2010, at

Re: [OMPI users] MPI_Send doesn't work if the data >= 2GB

2010-12-07 Thread Bowen Zhou
On 12/05/2010 10:13 PM, hi, I met a question recently when I tested the MPI_send and MPI_Recv functions. When I run the following codes, the processes hanged and I found there was not data transmission in my network at all. BTW: I finished this test on two X86-64 computers with 16GB memory

Re: [OMPI users] glut display 'occasionally' opens

2010-12-07 Thread brad baker
Thanks for your responses! I'm at home today so I can't actually do any tests to 'see' if anything works. But I logged in remotely and I did as Ralph suggested and ran env as my app. No process returned a value for DISPLAY. Then I made a small program that calls getenv("DISPLAY") to run with

Re: [OMPI users] glut display 'occasionally' opens

2010-12-07 Thread Jeff Squyres
Are you using ssh to launch OMPI between your nodes? (i.e., is mpirun using ssh under the covers to launch on remote nodes) If so, you might want to just set OMPI to use "ssh -X", which sets up SSH tunneled X forwarding, and therefore it sets DISPLAY for you properly on all the remote nodes

Re: [OMPI users] Scalability issue

2010-12-07 Thread Jeff Squyres
On Dec 7, 2010, at 8:33 AM, Gus Correa wrote: > Did I understand you right? > > Are you saying that one can effectively double the counting > capability (i.e. the "count" parameters in MPI calls) of OpenMPI > by compiling it with 8-byte integer flags? Yes and no. If you increase the size of

Re: [OMPI users] Scalability issue

2010-12-07 Thread Gus Correa
Hi Jeff Did I understand you right? Are you saying that one can effectively double the counting capability (i.e. the "count" parameters in MPI calls) of OpenMPI by compiling it with 8-byte integer flags? And long as one consistently uses the same flags to compile the application, everything

Re: [OMPI users] Segmentation fault in mca_pml_ob1.so

2010-12-07 Thread Grzegorz Maj
I recompiled MPI with -g, but it didn't solve the problem. Two things that have changed are: buf in PMPI_Recv is no longer of value 0 and backtrace in gdb shows more functions (eg. mca_pml_ob1_recv_frag_callback_put as #1). As you recommended, I will try to walk up the stack, but it's not so easy

Re: [OMPI users] Scalability issue

2010-12-07 Thread Jeff Squyres
It is always a good idea to have your application's sizeof(INTEGER) match the MPI's sizeof(INTEGER). Having them mismatch is a recipe for trouble. Meaning: if you're compiling your app with -make-integer-be-8-bytes, then you should configure/build Open MPI with that same flag. I'm thinking

Re: [OMPI users] Interaction with boost::asio

2010-12-07 Thread Ralph Castain
You might want to ask the boost people - we wouldn't have any idea what asio is or does. On Dec 7, 2010, at 6:06 AM, Hannes Brandstätter-Müller wrote: > Hello! > > I am using OpenMPI in combination with the boost libraries (especially > boost::asio) and came across a weird interaction. When I

Re: [OMPI users] Segmentation fault in mca_pml_ob1.so

2010-12-07 Thread Terry Dontje
I am not sure this has anything to do with your problem but if you look at the stack entry for PMPI_Recv I noticed the buf has a value of 0. Shouldn't that be an address? Does your code fail if the MPI library is built with -g? If it does fail the same way, the next step I would do would be

Re: [OMPI users] Segmentation fault in mca_pml_ob1.so

2010-12-07 Thread Grzegorz Maj
Some update on this issue. I've attached gdb to the crashing application and I got: - Program received signal SIGSEGV, Segmentation fault. mca_pml_ob1_send_request_put (sendreq=0x130c480, btl=0xc49850, hdr=0xd10e60) at pml_ob1_sendreq.c:1231 1231pml_ob1_sendreq.c: No such file or

Re: [OMPI users] MPI_Send doesn't work if the data >= 2GB

2010-12-07 Thread 孟宪军
Hi Gus Correa First of all, thanks for your suggestions. 1) The malloc function do return a non_NULL pointer. 2) I didn't tried the MPI_Isend function, actually, The really function I need to use is MPI_Allgatherv(). When I used it, I found this function didn't work when the the data >= 2GB,