[OMPI users] openmpi 2.1 large messages

2016-09-29 Thread Marlborough, Rick
Folks; I am attempting to set up a task that sends large messages via MPI_Bcast api. I am finding that small message work ok, anything less then 8000 bytes. Anything more than this then the whole scenario hangs with most of the worker processes pegged at 100% cpu usage. Tried

Re: [OMPI users] problems with client server scenario using MPI_Comm_connect

2016-10-04 Thread Marlborough, Rick
Gouaillardet Sent: Tuesday, October 04, 2016 7:13 AM To: Open MPI Users Subject: Re: [OMPI users] problems with client server scenario using MPI_Comm_connect Rick, I do not think ompi_server is required here. Can you please post a trimmed version of your client and server, and your two mpirun c

Re: [OMPI users] problems with client server scenario using MPI_Comm_connect

2016-10-04 Thread Marlborough, Rick
to try a nightly snapshot of the v2.0.x branch Cheers, Gilles On Tuesday, October 4, 2016, Marlborough, Rick <rmarlboro...@aaccorp.com<mailto:rmarlboro...@aaccorp.com>> wrote: Gilles; Here is the client side code. The start command is “mpirun –n 1 client 10” wher

[OMPI users] problems with client server scenario using MPI_Comm_connect

2016-10-03 Thread Marlborough, Rick
Folks; I have been trying to get a test case up and running using a client server scenario with a server waiting on MPI_Comm_accept and the client trying to connect via MPI_Comm_connect. The port value is written to a file. The client opens the file and reads the port value. I

[OMPI users] communicator with MPI_Reduce

2016-10-19 Thread Marlborough, Rick
Designation: Non-Export Controlled Content Folks; My MPI_COMM_WORLD size is 2000. I have created a communicator based on a small subset. I have used this communicator in MPI_Barrier calls and it seems to work fine. Now I want to use this in MPI_Reduce. When I do this I get

[OMPI users] clarity on Comm_connect

2016-10-12 Thread Marlborough, Rick
Designation: Non-Export Controlled Content Folks; Trying to do an MPI_Lookup_name. The call is surrounded by a try catch block. Even with the try catch block the calling process will still abort if the publishing process has not published the name. Is there a way to

Re: [OMPI users] clarity on Comm_connect

2016-10-12 Thread Marlborough, Rick
works! The sensors and proxies are all spawned in 2 batches using Comm_spawn_multiple. Error message below. Is there some configuration to enable this? [cid:image001.png@01D224B0.446AA710] 3.1.1001 From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Marlborough, Rick Sent

Re: [OMPI users] clarity on Comm_connect

2016-10-12 Thread Marlborough, Rick
@01D224B2.985D36B0] 3.1.1001 From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Marlborough, Rick Sent: Wednesday, October 12, 2016 5:44 PM To: Open MPI Users Subject: Re: [OMPI users] clarity on Comm_connect Designation: Non-Export Controlled Content ...forgot to mention... I have

[OMPI users] communications groups

2016-10-14 Thread Marlborough, Rick
Designation: Non-Export Controlled Content Folks; I have the following code setup. The sensorList is an array of ints of size 1. The value it contains is 1. My comm world size is 5. The call to MPI_Barrier fails every time with error "invalid communicator". This code is pretty

Re: [OMPI users] communications groups

2016-10-17 Thread Marlborough, Rick
communicator, which might explain the "invalid communicator" error you are getting. George. On Fri, Oct 14, 2016 at 5:33 PM, Marlborough, Rick <rmarlboro...@aaccorp.com<mailto:rmarlboro...@aaccorp.com>> wrote: Designation: Non-Export Controlled Content Folks;

Re: [OMPI users] communications groups

2016-10-17 Thread Marlborough, Rick
s, Gilles On Monday, October 17, 2016, Marlborough, Rick <rmarlboro...@aaccorp.com<mailto:rmarlboro...@aaccorp.com>> wrote: Designation: Non-Export Controlled Content George; Thanks for your response. Your second sentence is a little confusing. If my world group is P0,P1, vis

Re: [OMPI users] communications groups

2016-10-17 Thread Marlborough, Rick
and proxies in the right group Does that fit your needs ? If yes, then keep in mind sensorComm is MPI_COMM_NULL on the proxy tasks, proxyComm is MPI_COMM_NULL on the sensor tasks, and controlComm is MPI_COMM_NULL on the dispatcher. Cheers, Gilles On Monday, October 17, 2016, Marlborough, Rick