Hello,

I have some problems running MPI on my heterogeneous cluster. More
precisley i got segmentation fault when sending a large array (about
10000) of double from a i686 machine to a x86_64 machine. It does not
happen with small array. Here is the send/recv code source (complete
source is in attached file) :
========code ================
    if (me == 0 ) {
        for (int pe=1; pe<nprocs; pe++)
        {
                printf("Receiving from proc %d : ",pe); fflush(stdout);
            d=(double *)malloc(sizeof(double)*n);
            MPI_Recv(d,n,MPI_DOUBLE,pe,999,MPI_COMM_WORLD,&status);
            printf("OK\n"); fflush(stdout);
        }
        printf("All done.\n");
    }
    else {
      d=(double *)malloc(sizeof(double)*n);
      MPI_Send(d,n,MPI_DOUBLE,0,999,MPI_COMM_WORLD);
    }
======== code ================

I got segmentation fault with n=10000 but no error with n=1000
I have 2 machines :
sbtn155 : Intel Xeon,         x86_64
sbtn211 : Intel Pentium 4, i686

The code is compiled in x86_64 and i686 machine, using OpenMPI 1.4.1,
installed in /tmp/openmpi :
[mhtrinh@sbtn211 heterogenous]$ make hetero
gcc -Wall -I. -std=c99 -O3 -I/tmp/openmpi/include -c hetero.c -o hetero.i686.o
/tmp/openmpi/bin/mpicc -Wall -I. -std=c99 -O3 -I/tmp/openmpi/include
hetero.i686.o -o hetero.i686 -lm

[mhtrinh@sbtn155 heterogenous]$ make hetero
gcc -Wall -I. -std=c99 -O3 -I/tmp/openmpi/include -c hetero.c -o hetero.x86_64.o
/tmp/openmpi/bin/mpicc -Wall -I. -std=c99 -O3 -I/tmp/openmpi/include
hetero.x86_64.o -o hetero.x86_64 -lm

I run with the code using appfile and got thoses error :
$ cat appfile
--host sbtn155 -np 1 hetero.x86_64
--host sbtn155 -np 1 hetero.x86_64
--host sbtn211 -np 1 hetero.i686

$ mpirun -hetero --app appfile
Input array length :
10000
Receiving from proc 1 : OK
Receiving from proc 2 : [sbtn155:26386] *** Process received signal ***
[sbtn155:26386] Signal: Segmentation fault (11)
[sbtn155:26386] Signal code: Address not mapped (1)
[sbtn155:26386] Failing at address: 0x200627bd8
[sbtn155:26386] [ 0] /lib64/libpthread.so.0 [0x3fa4e0e540]
[sbtn155:26386] [ 1] /tmp/openmpi/lib/openmpi/mca_pml_ob1.so [0x2aaaad8d7908]
[sbtn155:26386] [ 2] /tmp/openmpi/lib/openmpi/mca_btl_tcp.so [0x2aaaae2fc6e3]
[sbtn155:26386] [ 3] /tmp/openmpi/lib/libopen-pal.so.0 [0x2aaaaafe39db]
[sbtn155:26386] [ 4]
/tmp/openmpi/lib/libopen-pal.so.0(opal_progress+0x9e) [0x2aaaaafd8b9e]
[sbtn155:26386] [ 5] /tmp/openmpi/lib/openmpi/mca_pml_ob1.so [0x2aaaad8d4b25]
[sbtn155:26386] [ 6] /tmp/openmpi/lib/libmpi.so.0(MPI_Recv+0x13b)
[0x2aaaaab30f9b]
[sbtn155:26386] [ 7] hetero.x86_64(main+0xde) [0x400cbe]
[sbtn155:26386] [ 8] /lib64/libc.so.6(__libc_start_main+0xf4) [0x3fa421e074]
[sbtn155:26386] [ 9] hetero.x86_64 [0x400b29]
[sbtn155:26386] *** End of error message ***
--------------------------------------------------------------------------
mpirun noticed that process rank 0 with PID 26386 on node sbtn155
exited on signal 11 (Segmentation fault).
--------------------------------------------------------------------------

Am I missing an option in order to run in heterogenous cluster ?
MPI_Send/Recv have limit array size when using heterogeneous cluster ?
Thanks for your help. Regards

-- 
============================================
   M. TRINH Minh Hieu
   CEA, IBEB, SBTN/LIRM,
   F-30207 Bagnols-sur-Cèze, FRANCE
============================================

Attachment: hetero.c.bz2
Description: BZip2 compressed data

Reply via email to