Re: [OMPI users] Possible buffer overflow on Recv rank

2019-03-27 Thread Gilles Gouaillardet
Carlos,

can you post a trimmed version of your code that evidences the issue ?

Keep in mind that if you want to write MPI code that is correct with respect to 
the standard, you should assume MPI_Send() might block until a matching receive 
is posted.

Cheers,

Gilles 

Sent from my iPod

> On Mar 27, 2019, at 20:46, carlos aguni  wrote:
> 
> Not "MPI_Send from 0"..
> MPI_Send from 1 to 0
> MPI_Send from 7 to 0
> And so on..
> 
>> On Wed, Mar 27, 2019, 8:43 AM carlos aguni  wrote:
>> Hi all.
>> 
>> I've an MPI application in which at one moment one rank receives a slice of 
>> an array from the other nodes.
>> Thing is that my application hangs there.
>> 
>> One thing I could get from printint out logs are:
>> (Rank 0) Starts MPI_Recv from source 4
>> But then it receives:
>> MPI_Send from 0
>> MPI_Send from 1
>> ... From 10
>> ... From 7
>> ... From 6
>> 
>> Then at one neither of them are responding.
>> The message is a double array type of size 100.000.
>> Later it would receive the message from 4.
>> 
>> So i assume the buffer on the Recv side is overflowing. 
>> 
>> Few tests:
>> - Using smaller array size works
>> - alreay tried using isend. Irecv. Bsend. And the ranks still get stuck.
>> 
>> So that leaves me to a few questions rather than how to solve this issue:
>> - how can i know the size of mpi's interbal buffer?
>> - how would one debug this?
>> 
>> Regards,
>> Carlos.
> ___
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Re: [OMPI users] Possible buffer overflow on Recv rank

2019-03-27 Thread carlos aguni
Not "MPI_Send from 0"..
MPI_Send from 1 to 0
MPI_Send from 7 to 0
And so on..

On Wed, Mar 27, 2019, 8:43 AM carlos aguni  wrote:

> Hi all.
>
> I've an MPI application in which at one moment one rank receives a slice
> of an array from the other nodes.
> Thing is that my application hangs there.
>
> One thing I could get from printint out logs are:
> (Rank 0) Starts MPI_Recv from source 4
> But then it receives:
> MPI_Send from 0
> MPI_Send from 1
> ... From 10
> ... From 7
> ... From 6
>
> Then at one neither of them are responding.
> The message is a double array type of size 100.000.
> Later it would receive the message from 4.
>
> So i assume the buffer on the Recv side is overflowing.
>
> Few tests:
> - Using smaller array size works
> - alreay tried using isend. Irecv. Bsend. And the ranks still get stuck.
>
> So that leaves me to a few questions rather than how to solve this issue:
> - how can i know the size of mpi's interbal buffer?
> - how would one debug this?
>
> Regards,
> Carlos.
>
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

[OMPI users] Possible buffer overflow on Recv rank

2019-03-27 Thread carlos aguni
Hi all.

I've an MPI application in which at one moment one rank receives a slice of
an array from the other nodes.
Thing is that my application hangs there.

One thing I could get from printint out logs are:
(Rank 0) Starts MPI_Recv from source 4
But then it receives:
MPI_Send from 0
MPI_Send from 1
... From 10
... From 7
... From 6

Then at one neither of them are responding.
The message is a double array type of size 100.000.
Later it would receive the message from 4.

So i assume the buffer on the Recv side is overflowing.

Few tests:
- Using smaller array size works
- alreay tried using isend. Irecv. Bsend. And the ranks still get stuck.

So that leaves me to a few questions rather than how to solve this issue:
- how can i know the size of mpi's interbal buffer?
- how would one debug this?

Regards,
Carlos.
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users