+1.  I'm afraid I don't know why offhand there would be such differences.  I'm 
thinking that you'll need to dive a little deeper to figure it out; sorry.  :-(


On Jan 16, 2011, at 10:54 AM, Shamis, Pavel wrote:

> Well, then I would suspect rdmacm vs oob QP configuration. They supposed to 
> be the same, but  probably it's some bug there, and  somehow rdmacm QP tuning 
> different from oob, it is potential source cause for the performance 
> differences that you see.
> 
> Regards,
> Pavel (Pasha) Shamis
> ---
> Application Performance Tools Group
> Computer Science and Math Division
> Oak Ridge National Laboratory
> 
> 
> 
> 
> 
> 
> On Jan 16, 2011, at 4:12 AM, Doron Shoham wrote:
> 
>> Hi,
>> 
>> The gather hangs only in liner_sync algorithm but works with
>> basic_linear and binomial algorithms.
>> The gather algorithm is choosen dynamiclly depanding on block size and
>> communicator size.
>> So, in the beginning, binomial algorithm is chosen (communicator size
>> is larger then 60).
>> When increasing the message size, the liner_sync algorithm is chosen
>> (with small_segment_size).
>> When debugging on the cluster I saw that the linear_sync function is
>> called in endless loop with segment size of 1024.
>> This explain why hang occure in the middle of the run.
>> 
>> I still don't understand why does RDMACM solve it or what causes
>> liner_sync hangs.
>> 
>> Again, in 1.5 it doesn't hang (maybe timing is different?).
>> I'm still trying to understand what are the diffrences in those areas
>> between 1.4.3 and 1.5
>> 
>> 
>> BTW,
>> Choosing RDMACM fixes hangs and performance issues in all collective 
>> operations.
>> 
>> Thanks,
>> Doron
>> 
>> 
>> On Thu, Jan 13, 2011 at 9:44 PM, Shamis, Pavel <sham...@ornl.gov> wrote:
>>> RDMACM creates the same QPs with the same tunings as OOB, so I don't see 
>>> how CPC may effect on performance.
>>> 
>>> Pavel (Pasha) Shamis
>>> ---
>>> Application Performance Tools Group
>>> Computer Science and Math Division
>>> Oak Ridge National Laboratory
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> On Jan 13, 2011, at 2:15 PM, Jeff Squyres wrote:
>>> 
>>>> +1 on what Pasha said -- if using rdmacm fixes the problem, then there's 
>>>> something else nefarious going on...
>>>> 
>>>> You might want to check padb with your hangs to see where all the 
>>>> processes are hung to see if anything obvious jumps out.  I'd be surprised 
>>>> if there's a bug in the oob cpc; it's been around for a long, long time; 
>>>> it should be pretty stable.
>>>> 
>>>> Do we create QP's differently between oob and rdmacm, such that perhaps 
>>>> they are "better" (maybe better routed, or using a different SL, or ...) 
>>>> when created via rdmacm?
>>>> 
>>>> 
>>>> On Jan 12, 2011, at 12:12 PM, Shamis, Pavel wrote:
>>>> 
>>>>> RDMACM or OOB can not effect on performance of this benchmark, since they 
>>>>> are not involved in communication. So I'm not sure that the performance 
>>>>> changes that you see are related to connection manager changes.
>>>>> About oob - I'm not aware about hangs issue there, the code is very-very 
>>>>> old, we did not touch it for a long time.
>>>>> 
>>>>> Regards,
>>>>> 
>>>>> Pavel (Pasha) Shamis
>>>>> ---
>>>>> Application Performance Tools Group
>>>>> Computer Science and Math Division
>>>>> Oak Ridge National Laboratory
>>>>> Email: sham...@ornl.gov
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> On Jan 12, 2011, at 8:45 AM, Doron Shoham wrote:
>>>>> 
>>>>>> Hi,
>>>>>> 
>>>>>> For the first problem, I can see that when using rdmacm as openib oob
>>>>>> I get much better performence results (and no hangs!).
>>>>>> 
>>>>>> mpirun -display-map -np 64 -machinefile voltairenodes -mca btl
>>>>>> sm,self,openib -mca btl_openib_connect_rdmacm_priority 100
>>>>>> imb/src/IMB-MPI1 gather -npmin 64
>>>>>> 
>>>>>> 
>>>>>> #bytes      #repetitions        t_min[usec]     t_max[usec]     
>>>>>> t_avg[usec]
>>>>>> 
>>>>>> 0       1000        0.04        0.05        0.05
>>>>>> 
>>>>>> 1       1000        19.64       19.69       19.67
>>>>>> 
>>>>>> 2       1000        19.97       20.02       19.99
>>>>>> 
>>>>>> 4       1000        21.86       21.96       21.89
>>>>>> 
>>>>>> 8       1000        22.87       22.94       22.90
>>>>>> 
>>>>>> 16      1000        24.71       24.80       24.76
>>>>>> 
>>>>>> 32      1000        27.23       27.32       27.27
>>>>>> 
>>>>>> 64      1000        30.96       31.06       31.01
>>>>>> 
>>>>>> 128     1000        36.96       37.08       37.02
>>>>>> 
>>>>>> 256     1000        42.64       42.79       42.72
>>>>>> 
>>>>>> 512     1000        60.32       60.59       60.46
>>>>>> 
>>>>>> 1024    1000        82.44       82.74       82.59
>>>>>> 
>>>>>> 2048    1000        497.66      499.62      498.70
>>>>>> 
>>>>>> 4096    1000        684.15      686.47      685.33
>>>>>> 
>>>>>> 8192    519         544.07      546.68      545.85
>>>>>> 
>>>>>> 16384   519         653.20      656.23      655.27
>>>>>> 
>>>>>> 32768   519         704.48      707.55      706.60
>>>>>> 
>>>>>> 65536   519         918.00      922.12      920.86
>>>>>> 
>>>>>> 131072  320         2414.08     2422.17     2418.20
>>>>>> 
>>>>>> 262144  160         4198.25     4227.58     4213.19
>>>>>> 
>>>>>> 524288  80          7333.04     7503.99     7438.18
>>>>>> 
>>>>>> 1048576 40          13692.60    14150.20    13948.75
>>>>>> 
>>>>>> 2097152 20          30377.34    32679.15    31779.86
>>>>>> 
>>>>>> 4194304 10          61416.70    71012.50    68380.04
>>>>>> 
>>>>>> How can the oob cause the hang? Isn't it only used to bring up the 
>>>>>> connection?
>>>>>> Does the oob has any part of the connections were made?
>>>>>> 
>>>>>> Thanks,
>>>>>> Dororn
>>>>>> 
>>>>>> On Tue, Jan 11, 2011 at 2:58 PM, Doron Shoham <doron.o...@gmail.com> 
>>>>>> wrote:
>>>>>>> 
>>>>>>> Hi
>>>>>>> 
>>>>>>> All machines on the setup are IDataPlex with Nehalem 12 cores per node, 
>>>>>>> 24GB  memory.
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> ·         Problem 1 – OMPI 1.4.3 hangs in gather:
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> I’m trying to run IMB and gather operation with OMPI 1.4.3 (Vanilla).
>>>>>>> 
>>>>>>> It happens when np >= 64 and message size exceed 4k:
>>>>>>> 
>>>>>>> mpirun -np 64 -machinefile voltairenodes -mca btl sm,self,openib  
>>>>>>> imb/src-1.4.2/IMB-MPI1 gather –npmin 64
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> voltairenodes consists of 64 machines.
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> #----------------------------------------------------------------
>>>>>>> 
>>>>>>> # Benchmarking Gather
>>>>>>> 
>>>>>>> # #processes = 64
>>>>>>> 
>>>>>>> #----------------------------------------------------------------
>>>>>>> 
>>>>>>>  #bytes #repetitions  t_min[usec]  t_max[usec]  t_avg[usec]
>>>>>>> 
>>>>>>>       0         1000         0.02         0.02         0.02
>>>>>>> 
>>>>>>>       1          331        14.02        14.16        14.09
>>>>>>> 
>>>>>>>       2          331        12.87        13.08        12.93
>>>>>>> 
>>>>>>>       4          331        14.29        14.43        14.34
>>>>>>> 
>>>>>>>       8          331        16.03        16.20        16.11
>>>>>>> 
>>>>>>>      16          331        17.54        17.74        17.64
>>>>>>> 
>>>>>>>      32          331        20.49        20.62        20.53
>>>>>>> 
>>>>>>>      64          331        23.57        23.84        23.70
>>>>>>> 
>>>>>>>     128          331        28.02        28.35        28.18
>>>>>>> 
>>>>>>>     256          331        34.78        34.88        34.80
>>>>>>> 
>>>>>>>     512          331        46.34        46.91        46.60
>>>>>>> 
>>>>>>>    1024          331        63.96        64.71        64.33
>>>>>>> 
>>>>>>>    2048          331       460.67       465.74       463.18
>>>>>>> 
>>>>>>>    4096          331       637.33       643.99       640.75
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> This the padb output:
>>>>>>> 
>>>>>>> padb –A –x –Ormgr=mpirun –tree:
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> =~=~=~=~=~=~=~=~=~=~=~= PuTTY log 2011.01.06 14:33:17 
>>>>>>> =~=~=~=~=~=~=~=~=~=~=~=
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> Warning, remote process state differs across ranks
>>>>>>> 
>>>>>>> state : ranks
>>>>>>> 
>>>>>>> R (running) : 
>>>>>>> [1,3-6,8,10-13,16-20,23-28,30-32,34-42,44-45,47-49,51-53,56-59,61-63]
>>>>>>> 
>>>>>>> S (sleeping) : [0,2,7,9,14-15,21-22,29,33,43,46,50,54-55,60]
>>>>>>> 
>>>>>>> Stack trace(s) for thread: 1
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> [0-63] (64 processes)
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> main() at ?:?
>>>>>>> 
>>>>>>> IMB_init_buffers_iter() at ?:?
>>>>>>> 
>>>>>>> IMB_gather() at ?:?
>>>>>>> 
>>>>>>> PMPI_Gather() at pgather.c:175
>>>>>>> 
>>>>>>>   mca_coll_sync_gather() at coll_sync_gather.c:46
>>>>>>> 
>>>>>>>     ompi_coll_tuned_gather_intra_dec_fixed() at 
>>>>>>> coll_tuned_decision_fixed.c:714
>>>>>>> 
>>>>>>>       -----------------
>>>>>>> 
>>>>>>>       [0,3-63] (62 processes)
>>>>>>> 
>>>>>>>       -----------------
>>>>>>> 
>>>>>>>       ompi_coll_tuned_gather_intra_linear_sync() at 
>>>>>>> coll_tuned_gather.c:248
>>>>>>> 
>>>>>>>         mca_pml_ob1_recv() at pml_ob1_irecv.c:104
>>>>>>> 
>>>>>>>           ompi_request_wait_completion() at 
>>>>>>> ../../../../ompi/request/request.h:375
>>>>>>> 
>>>>>>>             opal_condition_wait() at 
>>>>>>> ../../../../opal/threads/condition.h:99
>>>>>>> 
>>>>>>>       -----------------
>>>>>>> 
>>>>>>>       [1] (1 processes)
>>>>>>> 
>>>>>>>       -----------------
>>>>>>> 
>>>>>>>       ompi_coll_tuned_gather_intra_linear_sync() at 
>>>>>>> coll_tuned_gather.c:302
>>>>>>> 
>>>>>>>         mca_pml_ob1_send() at pml_ob1_isend.c:125
>>>>>>> 
>>>>>>>           ompi_request_wait_completion() at 
>>>>>>> ../../../../ompi/request/request.h:375
>>>>>>> 
>>>>>>>             opal_condition_wait() at 
>>>>>>> ../../../../opal/threads/condition.h:99
>>>>>>> 
>>>>>>>       -----------------
>>>>>>> 
>>>>>>>       [2] (1 processes)
>>>>>>> 
>>>>>>>       -----------------
>>>>>>> 
>>>>>>>       ompi_coll_tuned_gather_intra_linear_sync() at 
>>>>>>> coll_tuned_gather.c:315
>>>>>>> 
>>>>>>>         ompi_request_default_wait() at request/req_wait.c:37
>>>>>>> 
>>>>>>>           ompi_request_wait_completion() at 
>>>>>>> ../ompi/request/request.h:375
>>>>>>> 
>>>>>>>             opal_condition_wait() at ../opal/threads/condition.h:99
>>>>>>> 
>>>>>>> Stack trace(s) for thread: 2
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> [0-63] (64 processes)
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> start_thread() at ?:?
>>>>>>> 
>>>>>>> btl_openib_async_thread() at btl_openib_async.c:344
>>>>>>> 
>>>>>>> poll() at ?:?
>>>>>>> 
>>>>>>> Stack trace(s) for thread: 3
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> [0-63] (64 processes)
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> start_thread() at ?:?
>>>>>>> 
>>>>>>> service_thread_start() at btl_openib_fd.c:427
>>>>>>> 
>>>>>>> select() at ?:?
>>>>>>> 
>>>>>>> -bash-3.2$
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> When running again padb after couple of minutes, I can see that the 
>>>>>>> total number of processes remain in the same position but
>>>>>>> 
>>>>>>> different processes are at different positions.
>>>>>>> 
>>>>>>> For example, this is the diff between two padb outputs:
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> Warning, remote process state differs across ranks
>>>>>>> 
>>>>>>> state : ranks
>>>>>>> 
>>>>>>> -R (running) : [0,2-4,6-13,16-18,20-21,28-31,33-36,38-56,58,60,62-63]
>>>>>>> 
>>>>>>> -S (sleeping) : [1,5,14-15,19,22-27,32,37,57,59,61]
>>>>>>> 
>>>>>>> +R (running) : [2,5-14,16-23,25,28-40,42-48,50-51,53-58,61,63]
>>>>>>> 
>>>>>>> +S (sleeping) : [0-1,3-4,15,24,26-27,41,49,52,59-60,62]
>>>>>>> 
>>>>>>> Stack trace(s) for thread: 1
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> [0-63] (64 processes)
>>>>>>> 
>>>>>>> @@ -13,21 +13,21 @@
>>>>>>> 
>>>>>>> mca_coll_sync_gather() at coll_sync_gather.c:46
>>>>>>> 
>>>>>>> ompi_coll_tuned_gather_intra_dec_fixed() at 
>>>>>>> coll_tuned_decision_fixed.c:714
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> - [0,3-63] (62 processes)
>>>>>>> 
>>>>>>> + [0-5,8-63] (62 processes)
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> ompi_coll_tuned_gather_intra_linear_sync() at coll_tuned_gather.c:248
>>>>>>> 
>>>>>>> mca_pml_ob1_recv() at pml_ob1_irecv.c:104
>>>>>>> 
>>>>>>> ompi_request_wait_completion() at ../../../../ompi/request/request.h:375
>>>>>>> 
>>>>>>> opal_condition_wait() at ../../../../opal/threads/condition.h:99
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> - [1] (1 processes)
>>>>>>> 
>>>>>>> + [6] (1 processes)
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> ompi_coll_tuned_gather_intra_linear_sync() at coll_tuned_gather.c:302
>>>>>>> 
>>>>>>> mca_pml_ob1_send() at pml_ob1_isend.c:125
>>>>>>> 
>>>>>>> ompi_request_wait_completion() at ../../../../ompi/request/request.h:375
>>>>>>> 
>>>>>>> opal_condition_wait() at ../../../../opal/threads/condition.h:99
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> - [2] (1 processes)
>>>>>>> 
>>>>>>> + [7] (1 processes)
>>>>>>> 
>>>>>>> -----------------
>>>>>>> 
>>>>>>> ompi_coll_tuned_gather_intra_linear_sync() at coll_tuned_gather.c:315
>>>>>>> 
>>>>>>> ompi_request_default_wait() at request/req_wait.c:37
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> Choosing different gather algorithm seems to bypass the hang.
>>>>>>> 
>>>>>>> I’ve used the following mca parameters:
>>>>>>> 
>>>>>>> --mca coll_tuned_use_dynamic_rules 1
>>>>>>> 
>>>>>>> --mca coll_tuned_gather_algorithm 1
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> Actually, both dec_fixed and basic_linear works while binomial and 
>>>>>>> linear_sync doesn’t.
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> With OMPI 1.5 it doesn’t hangs (with all gather algorithms) and it much 
>>>>>>> faster (the number of repetitions is much higher):
>>>>>>> 
>>>>>>> #----------------------------------------------------------------
>>>>>>> 
>>>>>>> # Benchmarking Gather
>>>>>>> 
>>>>>>> # #processes = 64
>>>>>>> 
>>>>>>> #----------------------------------------------------------------
>>>>>>> 
>>>>>>>  #bytes #repetitions  t_min[usec]  t_max[usec]  t_avg[usec]
>>>>>>> 
>>>>>>>       0         1000         0.02         0.03         0.02
>>>>>>> 
>>>>>>>       1         1000        18.50        18.55        18.53
>>>>>>> 
>>>>>>>       2         1000        18.17        18.25        18.22
>>>>>>> 
>>>>>>>       4         1000        19.04        19.10        19.07
>>>>>>> 
>>>>>>>       8         1000        19.60        19.67        19.64
>>>>>>> 
>>>>>>>      16         1000        21.39        21.47        21.43
>>>>>>> 
>>>>>>>      32         1000        24.83        24.91        24.87
>>>>>>> 
>>>>>>>      64         1000        27.35        27.45        27.40
>>>>>>> 
>>>>>>>     128         1000        33.23        33.34        33.29
>>>>>>> 
>>>>>>>     256         1000        41.24        41.39        41.32
>>>>>>> 
>>>>>>>     512         1000        52.62        52.81        52.71
>>>>>>> 
>>>>>>>    1024         1000        73.20        73.46        73.32
>>>>>>> 
>>>>>>>    2048         1000       416.36       418.04       417.22
>>>>>>> 
>>>>>>>    4096         1000       638.54       640.70       639.65
>>>>>>> 
>>>>>>>    8192         1000       506.26       506.97       506.63
>>>>>>> 
>>>>>>>   16384         1000       600.63       601.40       601.02
>>>>>>> 
>>>>>>>   32768         1000       639.52       640.34       639.93
>>>>>>> 
>>>>>>>   65536          640       914.22       916.02       915.13
>>>>>>> 
>>>>>>>  131072          320      2287.37      2295.18      2291.35
>>>>>>> 
>>>>>>>  262144          160      4041.36      4070.58      4056.27
>>>>>>> 
>>>>>>>  524288           80      7292.35      7463.27      7397.14
>>>>>>> 
>>>>>>> 1048576           40     13647.15     14107.15     13905.29
>>>>>>> 
>>>>>>> 2097152           20     30625.00     32635.45     31815.36
>>>>>>> 
>>>>>>> 4194304           10     63543.01     70987.49     68680.48
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> ·         Problem 2 – segmentation fault with OMPI 1.4.3/1.5 and IMB 
>>>>>>> gather np=768:
>>>>>>> 
>>>>>>> When trying to run the same command but with np=768 I get segmentation 
>>>>>>> fault:
>>>>>>> 
>>>>>>> openmpi-1.4.3/bin/mpirun -np 768 -machinefile voltairenodes -mca btl 
>>>>>>> sm,self,openib -mca coll_tuned_use_dynamic_rules 1 -mca 
>>>>>>> coll_tuned_gather_algorithm 1 imb/src/IMB-MPI1 gather -npmin 768 -mem 
>>>>>>> 1.6
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> This happens in OMPI 1.4.3 and 1.5
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> [compa163:20249] *** Process received signal ***
>>>>>>> 
>>>>>>> [compa163:20249] Signal: Segmentation fault (11)
>>>>>>> 
>>>>>>> [compa163:20249] Signal code: Address not mapped (1)
>>>>>>> 
>>>>>>> [compa163:20249] Failing at address: 0x2aab4a204000
>>>>>>> 
>>>>>>> [compa163:20249] [ 0] /lib64/libpthread.so.0 [0x366aa0e7c0]
>>>>>>> 
>>>>>>> [compa163:20249] [ 1] 
>>>>>>> /gpfs/asrc/home/voltaire/install//openmpi-1.4.3/lib/libmpi.so.0(ompi_convertor_unpack+0x15f)
>>>>>>>  [0x2b077882282e]
>>>>>>> 
>>>>>>> [compa163:20249] [ 2] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_pml_ob1.so
>>>>>>>  [0x2b077b9e1672]
>>>>>>> 
>>>>>>> [compa163:20249] [ 3] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_pml_ob1.so
>>>>>>>  [0x2b077b9dd0b6]
>>>>>>> 
>>>>>>> [compa163:20249] [ 4] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_btl_sm.so
>>>>>>>  [0x2b077c459d87]
>>>>>>> 
>>>>>>> [compa163:20249] [ 5] 
>>>>>>> /gpfs/asrc/home/voltaire/install//openmpi-1.4.3/lib/libopen-pal.so.0(opal_progress+0xbe)
>>>>>>>  [0x2b0778d845b8]
>>>>>>> 
>>>>>>> [compa163:20249] [ 6] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_pml_ob1.so
>>>>>>>  [0x2b077b9d6d62]
>>>>>>> 
>>>>>>> [compa163:20249] [ 7] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_pml_ob1.so
>>>>>>>  [0x2b077b9d6ba7]
>>>>>>> 
>>>>>>> [compa163:20249] [ 8] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_pml_ob1.so
>>>>>>>  [0x2b077b9d6a90]
>>>>>>> 
>>>>>>> [compa163:20249] [ 9] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_coll_tuned.so
>>>>>>>  [0x2b077d298dc5]
>>>>>>> 
>>>>>>> [compa163:20249] [10] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_coll_tuned.so
>>>>>>>  [0x2b077d2990d3]
>>>>>>> 
>>>>>>> [compa163:20249] [11] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_coll_tuned.so
>>>>>>>  [0x2b077d286e9b]
>>>>>>> 
>>>>>>> [compa163:20249] [12] 
>>>>>>> /gpfs/asrc/home/voltaire/install/openmpi-1.4.3/lib/openmpi/mca_coll_sync.so
>>>>>>>  [0x2b077d07e96c]
>>>>>>> 
>>>>>>> [compa163:20249] [13] 
>>>>>>> /gpfs/asrc/home/voltaire/install//openmpi-1.4.3/lib/libmpi.so.0(PMPI_Gather+0x55e)
>>>>>>>  [0x2b077883ec9a]
>>>>>>> 
>>>>>>> [compa163:20249] [14] imb/src/IMB-MPI1(IMB_gather+0xe8) [0x40a088]
>>>>>>> 
>>>>>>> [compa163:20249] [15] imb/src/IMB-MPI1(IMB_init_buffers_iter+0x28a) 
>>>>>>> [0x405baa]
>>>>>>> 
>>>>>>> [compa163:20249] [16] imb/src/IMB-MPI1(main+0x30f) [0x40362f]
>>>>>>> 
>>>>>>> [compa163:20249] [17] /lib64/libc.so.6(__libc_start_main+0xf4) 
>>>>>>> [0x3669e1d994]
>>>>>>> 
>>>>>>> [compa163:20249] [18] imb/src/IMB-MPI1 [0x403269]
>>>>>>> 
>>>>>>> [compa163:20249] *** End of error message ***
>>>>>>> 
>>>>>>> 
>>>>>>> Any ideas? More debuggin tips?
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Doron
>>>>>> 
>>>>>> _______________________________________________
>>>>>> devel mailing list
>>>>>> de...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> devel mailing list
>>>>> de...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>> 
>>>> 
>>>> --
>>>> Jeff Squyres
>>>> jsquy...@cisco.com
>>>> For corporate legal information go to:
>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>> 
>>>> 
>>>> _______________________________________________
>>>> devel mailing list
>>>> de...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> 
>>> 
>>> _______________________________________________
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> 
>> 
>> _______________________________________________
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> 
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to