Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-19 Thread Jeff Squyres
Whenever I have seen this problem, it has been because of a mismatch between 
mpirun and the back-end libraries that are linked against the executable.  For 
example, the app was compiled/linked against Open MPI a.b.c and either some 
other mpirun was used (e.g., from MPICH) or the mpirun from Open MPI x.y.z was 
used.



On Apr 18, 2010, at 2:28 PM, Ralph Castain wrote:

> Afraid I can't help you - I've never seen that behavior on any system, can't 
> replicate it anywhere, and have no idea what might cause it.
> 
> On Apr 18, 2010, at 9:15 AM, Mario Ogrizek wrote:
> 
>> It is a parallel tools platform for eclipse IDE, a plugin.
>> I dont think it is a source of problem.
>> 
>> The same thing is happening running it from shell. It has something to do 
>> with mapping or something else. Since it allways maps for job 0, what ever 
>> that means.
>> 
>> On Sun, Apr 18, 2010 at 4:50 PM, Ralph Castain  wrote:
>> Again, what is PTP?
>> 
>> I can't replicate this on any system we can access, so it may be something 
>> about this PTP thing.
>> 
>> On Apr 18, 2010, at 1:37 AM, Mario Ogrizek wrote:
>> 
>>> Ofcourse i checked that, i have all of this things, 
>>> I simplified the program, and its the same.
>>> Nothing gave me clue, except the more detailed writeout of the PTP.
>>> Here is the critical part of it:
>>> (1.2 one, this is correct)
>>> [Mario.local:05548]  Map for job: 1 Generated by mapping mode: byslot
>>> Starting vpid: 0Vpid range: 4   Num app_contexts: 1
>>> ...
>>> ...
>>> 
>>> (1.4 one)
>>> [Mario.local:05542]  Map for job: 0 Generated by mapping mode: byslot
>>> Starting vpid: 0Vpid range: 1   Num app_contexts: 1
>>> ...
>>> ...
>>> 
>>> Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to, 
>>> but hope it will give you some clues.
>>>  
>>> On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain  wrote:
>>> Just to check what is going on, why don't you remove that message passing 
>>> code and just
>>> 
>>> printf("Hello MPI World from process %d!", my_rank
>>> 
>>>  in each process? Much more direct - avoids any ambiguity.
>>> 
>>> Also, be certain that you compile this program for the specific OMPI 
>>> version you are running it under. OMPI is NOT binary compatible across 
>>> releases - you have to recompile the program for the specific release you 
>>> are going to use.
>>> 
>>> 
>>> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
>>> 
 Ofcourse, its the same program, wasnt recompiled for a week.
 
 
 #include 
 #include 
 #include "mpi.h"
 
 int main(int argc, char* argv[]){
int  my_rank; /* rank of process */
int  p;   /* number of processes */
int source;   /* rank of sender */
int dest; /* rank of receiver */
int tag=0;/* tag for messages */
char message[100];/* storage for message */
MPI_Status status ;   /* return status for receive */

/* start up MPI */

MPI_Init(, );
 
/* find out process rank */
MPI_Comm_rank(MPI_COMM_WORLD, _rank); 

 
/* find out number of processes */
MPI_Comm_size(MPI_COMM_WORLD, );
 

if (my_rank !=0){
/* create message */
sprintf(message, "Hello MPI World from process %d!", my_rank);
dest = 0;
/* use strlen+1 so that '\0' get transmitted */
MPI_Send(message, strlen(message)+1, MPI_CHAR,
   dest, tag, MPI_COMM_WORLD);
}
else{
printf("Hello MPI World From process 0: Num processes: %d\n",p);
for (source = 1; source < p; source++) {
MPI_Recv(message, 100, MPI_CHAR, source, tag,
  MPI_COMM_WORLD, );
printf("%s\n",message);
}
}
/* shut down MPI */
MPI_Finalize(); 


return 0;
 }
 
 I triplechecked:
 v1.2 output
 Hello MPI World From process 0: Num processes: 4
 Hello MPI World from process 1!
 Hello MPI World from process 2!
 Hello MPI World from process 3!
 
 v1.4 output:
 Hello MPI World From process 0: Num processes: 1
 Hello MPI World From process 0: Num processes: 1
 Hello MPI World From process 0: Num processes: 1
 Hello MPI World From process 0: Num processes: 1
 
 
 
 
 
 
 
 On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain  wrote:
 
 On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
 
> Hahaha, ok then that WAS silly! :D
> So there is no way to utilize both cores with mpi?
 
 We are using both cores - it is just that they are on the same node. 
 Unless told otherwise, the processes will use shared memory for 
 communication.
 
> 
> Ah well, I'll 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-18 Thread Ralph Castain
Afraid I can't help you - I've never seen that behavior on any system, can't 
replicate it anywhere, and have no idea what might cause it.

On Apr 18, 2010, at 9:15 AM, Mario Ogrizek wrote:

> It is a parallel tools platform for eclipse IDE, a plugin.
> I dont think it is a source of problem.
> 
> The same thing is happening running it from shell. It has something to do 
> with mapping or something else. Since it allways maps for job 0, what ever 
> that means.
> 
> On Sun, Apr 18, 2010 at 4:50 PM, Ralph Castain  wrote:
> Again, what is PTP?
> 
> I can't replicate this on any system we can access, so it may be something 
> about this PTP thing.
> 
> On Apr 18, 2010, at 1:37 AM, Mario Ogrizek wrote:
> 
>> Ofcourse i checked that, i have all of this things, 
>> I simplified the program, and its the same.
>> Nothing gave me clue, except the more detailed writeout of the PTP.
>> Here is the critical part of it:
>> (1.2 one, this is correct)
>> [Mario.local:05548]  Map for job: 1  Generated by mapping mode: byslot
>>  Starting vpid: 0Vpid range: 4   Num app_contexts: 1
>> ...
>> ...
>> 
>> (1.4 one)
>> [Mario.local:05542]  Map for job: 0  Generated by mapping mode: byslot
>>  Starting vpid: 0Vpid range: 1   Num app_contexts: 1
>> ...
>> ...
>> 
>> Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to, 
>> but hope it will give you some clues.
>>  
>> On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain  wrote:
>> Just to check what is going on, why don't you remove that message passing 
>> code and just
>> 
>> printf("Hello MPI World from process %d!", my_rank
>> 
>>  in each process? Much more direct - avoids any ambiguity.
>> 
>> Also, be certain that you compile this program for the specific OMPI version 
>> you are running it under. OMPI is NOT binary compatible across releases - 
>> you have to recompile the program for the specific release you are going to 
>> use.
>> 
>> 
>> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
>> 
>>> Ofcourse, its the same program, wasnt recompiled for a week.
>>> 
>>> 
>>> #include 
>>> #include 
>>> #include "mpi.h"
>>> 
>>> int main(int argc, char* argv[]){
>>> int  my_rank; /* rank of process */
>>> int  p;   /* number of processes */
>>> int source;   /* rank of sender */
>>> int dest; /* rank of receiver */
>>> int tag=0;/* tag for messages */
>>> char message[100];/* storage for message */
>>> MPI_Status status ;   /* return status for receive */
>>> 
>>> /* start up MPI */
>>> 
>>> MPI_Init(, );
>>> 
>>> /* find out process rank */
>>> MPI_Comm_rank(MPI_COMM_WORLD, _rank); 
>>> 
>>> 
>>> /* find out number of processes */
>>> MPI_Comm_size(MPI_COMM_WORLD, );
>>> 
>>> 
>>> if (my_rank !=0){
>>> /* create message */
>>> sprintf(message, "Hello MPI World from process %d!", my_rank);
>>> dest = 0;
>>> /* use strlen+1 so that '\0' get transmitted */
>>> MPI_Send(message, strlen(message)+1, MPI_CHAR,
>>>dest, tag, MPI_COMM_WORLD);
>>> }
>>> else{
>>> printf("Hello MPI World From process 0: Num processes: %d\n",p);
>>> for (source = 1; source < p; source++) {
>>> MPI_Recv(message, 100, MPI_CHAR, source, tag,
>>>   MPI_COMM_WORLD, );
>>> printf("%s\n",message);
>>> }
>>> }
>>> /* shut down MPI */
>>> MPI_Finalize(); 
>>> 
>>> 
>>> return 0;
>>> }
>>> 
>>> I triplechecked:
>>> v1.2 output
>>> Hello MPI World From process 0: Num processes: 4
>>> Hello MPI World from process 1!
>>> Hello MPI World from process 2!
>>> Hello MPI World from process 3!
>>> 
>>> v1.4 output:
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain  wrote:
>>> 
>>> On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
>>> 
 Hahaha, ok then that WAS silly! :D
 So there is no way to utilize both cores with mpi?
>>> 
>>> We are using both cores - it is just that they are on the same node. Unless 
>>> told otherwise, the processes will use shared memory for communication.
>>> 
 
 Ah well, I'll correct that.
 
 From console, im starting a job like this: mpirun -np 4 Program, where i 
 want to run a Program on 4 processors.
 I was just stumbled when i got same output 4 times, like there are 4 
 processes ranked 0.
 While with the old version of mpi (1.2) same execution would give 4 
 processes ranked 0..3.
>>> 
>>> And so you should - if not, then there is something wrong. No way mpirun 
>>> would start 4 processes ranked 0. How are you 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-18 Thread Mario Ogrizek
It is a parallel tools platform for eclipse IDE, a plugin.
I dont think it is a source of problem.

The same thing is happening running it from shell. It has something to do
with mapping or something else. Since it allways maps for job 0, what ever
that means.

On Sun, Apr 18, 2010 at 4:50 PM, Ralph Castain  wrote:

> Again, what is PTP?
>
> I can't replicate this on any system we can access, so it may be something
> about this PTP thing.
>
> On Apr 18, 2010, at 1:37 AM, Mario Ogrizek wrote:
>
> Ofcourse i checked that, i have all of this things,
> I simplified the program, and its the same.
> Nothing gave me clue, except the more detailed writeout of the PTP.
> Here is the critical part of it:
> (1.2 one, this is correct)
> [Mario.local:05548]  Map for job: 1 Generated by mapping mode: byslot
>   Starting vpid: 0 Vpid range: 4 Num app_contexts: 1
> ...
> ...
>
> (1.4 one)
> [Mario.local:05542]  Map for job: 0 Generated by mapping mode: byslot
>   Starting vpid: 0 Vpid range: 1 Num app_contexts: 1
> ...
> ...
>
> Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to,
> but hope it will give you some clues.
>
> On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain  wrote:
>
>> Just to check what is going on, why don't you remove that message passing
>> code and just
>>
>> printf("Hello MPI World from process %d!", my_rank
>>
>>  in each process? Much more direct - avoids any ambiguity.
>>
>> Also, be certain that you compile this program for the specific OMPI
>> version you are running it under. OMPI is NOT binary compatible across
>> releases - you have to recompile the program for the specific release you
>> are going to use.
>>
>>
>> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
>>
>> Ofcourse, its the same program, wasnt recompiled for a week.
>>
>>
>> #include 
>> #include 
>> #include "mpi.h"
>>
>> int main(int argc, char* argv[]){
>>  int  my_rank; /* rank of process */
>>  int  p;   /* number of processes */
>>  int source;   /* rank of sender */
>>  int dest; /* rank of receiver */
>>  int tag=0;/* tag for messages */
>>  char message[100];/* storage for message */
>>  MPI_Status status ;   /* return status for receive */
>>
>>  /* start up MPI */
>>
>>  MPI_Init(, );
>>
>> /* find out process rank */
>>  MPI_Comm_rank(MPI_COMM_WORLD, _rank);
>>
>>
>> /* find out number of processes */
>>  MPI_Comm_size(MPI_COMM_WORLD, );
>>
>>
>> if (my_rank !=0){
>>  /* create message */
>>  sprintf(message, "Hello MPI World from process %d!", my_rank);
>>  dest = 0;
>> /* use strlen+1 so that '\0' get transmitted */
>>  MPI_Send(message, strlen(message)+1, MPI_CHAR,
>>dest, tag, MPI_COMM_WORLD);
>>  }
>> else{
>>  printf("Hello MPI World From process 0: Num processes: %d\n",p);
>>  for (source = 1; source < p; source++) {
>>  MPI_Recv(message, 100, MPI_CHAR, source, tag,
>>   MPI_COMM_WORLD, );
>>  printf("%s\n",message);
>>  }
>> }
>>  /* shut down MPI */
>>  MPI_Finalize();
>>
>>
>>  return 0;
>> }
>>
>> I triplechecked:
>> v1.2 output
>> Hello MPI World From process 0: Num processes: 4
>> Hello MPI World from process 1!
>> Hello MPI World from process 2!
>> Hello MPI World from process 3!
>>
>> v1.4 output:
>>
>> Hello MPI World From process 0: Num processes: 1
>>
>> Hello MPI World From process 0: Num processes: 1
>>
>> Hello MPI World From process 0: Num processes: 1
>>
>> Hello MPI World From process 0: Num processes: 1
>>
>>
>>
>>
>>
>>
>>
>> On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain  wrote:
>>
>>>
>>> On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
>>>
>>> Hahaha, ok then that WAS silly! :D
>>> So there is no way to utilize both cores with mpi?
>>>
>>>
>>> We are using both cores - it is just that they are on the same node.
>>> Unless told otherwise, the processes will use shared memory for
>>> communication.
>>>
>>>
>>> Ah well, I'll correct that.
>>>
>>> From console, im starting a job like this: mpirun -np 4 Program, where i
>>> want to run a Program on 4 processors.
>>> I was just stumbled when i got same output 4 times, like there are 4
>>> processes ranked 0.
>>> While with the old version of mpi (1.2) same execution would give 4
>>> processes ranked 0..3.
>>>
>>>
>>> And so you should - if not, then there is something wrong. No way mpirun
>>> would start 4 processes ranked 0. How are you printing the rank? Are you
>>> sure you are getting it correctly?
>>>
>>>
>>>
>>> Hope you see my question.
>>>
>>> On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain  wrote:
>>>

 On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:

 I am new to mpi, so I'm sorry for any silly questions.

 My idea was to try to use dual core machine as two nodes. I have a
 limited access to a cluster, so this was just for "testing" purposes.
 My default hostfile contains usual comments and this two nodes:

 node0
 node1

 I thought that each processor is a 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-18 Thread Ralph Castain
Again, what is PTP?

I can't replicate this on any system we can access, so it may be something 
about this PTP thing.

On Apr 18, 2010, at 1:37 AM, Mario Ogrizek wrote:

> Ofcourse i checked that, i have all of this things, 
> I simplified the program, and its the same.
> Nothing gave me clue, except the more detailed writeout of the PTP.
> Here is the critical part of it:
> (1.2 one, this is correct)
> [Mario.local:05548]  Map for job: 1   Generated by mapping mode: byslot
>   Starting vpid: 0Vpid range: 4   Num app_contexts: 1
> ...
> ...
> 
> (1.4 one)
> [Mario.local:05542]  Map for job: 0   Generated by mapping mode: byslot
>   Starting vpid: 0Vpid range: 1   Num app_contexts: 1
> ...
> ...
> 
> Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to, but 
> hope it will give you some clues.
>  
> On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain  wrote:
> Just to check what is going on, why don't you remove that message passing 
> code and just
> 
> printf("Hello MPI World from process %d!", my_rank
> 
>  in each process? Much more direct - avoids any ambiguity.
> 
> Also, be certain that you compile this program for the specific OMPI version 
> you are running it under. OMPI is NOT binary compatible across releases - you 
> have to recompile the program for the specific release you are going to use.
> 
> 
> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
> 
>> Ofcourse, its the same program, wasnt recompiled for a week.
>> 
>> 
>> #include 
>> #include 
>> #include "mpi.h"
>> 
>> int main(int argc, char* argv[]){
>>  int  my_rank; /* rank of process */
>>  int  p;   /* number of processes */
>>  int source;   /* rank of sender */
>>  int dest; /* rank of receiver */
>>  int tag=0;/* tag for messages */
>>  char message[100];/* storage for message */
>>  MPI_Status status ;   /* return status for receive */
>>  
>>  /* start up MPI */
>>  
>>  MPI_Init(, );
>> 
>>  /* find out process rank */
>>  MPI_Comm_rank(MPI_COMM_WORLD, _rank); 
>>  
>> 
>>  /* find out number of processes */
>>  MPI_Comm_size(MPI_COMM_WORLD, );
>> 
>>  
>>  if (my_rank !=0){
>>  /* create message */
>>  sprintf(message, "Hello MPI World from process %d!", my_rank);
>>  dest = 0;
>>  /* use strlen+1 so that '\0' get transmitted */
>>  MPI_Send(message, strlen(message)+1, MPI_CHAR,
>> dest, tag, MPI_COMM_WORLD);
>>  }
>>  else{
>>  printf("Hello MPI World From process 0: Num processes: %d\n",p);
>>  for (source = 1; source < p; source++) {
>>  MPI_Recv(message, 100, MPI_CHAR, source, tag,
>>MPI_COMM_WORLD, );
>>  printf("%s\n",message);
>>  }
>>  }
>>  /* shut down MPI */
>>  MPI_Finalize(); 
>>  
>>  
>>  return 0;
>> }
>> 
>> I triplechecked:
>> v1.2 output
>> Hello MPI World From process 0: Num processes: 4
>> Hello MPI World from process 1!
>> Hello MPI World from process 2!
>> Hello MPI World from process 3!
>> 
>> v1.4 output:
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain  wrote:
>> 
>> On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
>> 
>>> Hahaha, ok then that WAS silly! :D
>>> So there is no way to utilize both cores with mpi?
>> 
>> We are using both cores - it is just that they are on the same node. Unless 
>> told otherwise, the processes will use shared memory for communication.
>> 
>>> 
>>> Ah well, I'll correct that.
>>> 
>>> From console, im starting a job like this: mpirun -np 4 Program, where i 
>>> want to run a Program on 4 processors.
>>> I was just stumbled when i got same output 4 times, like there are 4 
>>> processes ranked 0.
>>> While with the old version of mpi (1.2) same execution would give 4 
>>> processes ranked 0..3.
>> 
>> And so you should - if not, then there is something wrong. No way mpirun 
>> would start 4 processes ranked 0. How are you printing the rank? Are you 
>> sure you are getting it correctly?
>> 
>> 
>>> 
>>> Hope you see my question.
>>> 
>>> On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain  wrote:
>>> 
>>> On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:
>>> 
 I am new to mpi, so I'm sorry for any silly questions.
 
 My idea was to try to use dual core machine as two nodes. I have a limited 
 access to a cluster, so this was just for "testing" purposes.
 My default hostfile contains usual comments and this two nodes:
 
> node0
> node1
 I thought that each processor is a node for MPI purpose.
>>> 
>>> I'm afraid 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-18 Thread Mario Ogrizek
Ofcourse i checked that, i have all of this things,
I simplified the program, and its the same.
Nothing gave me clue, except the more detailed writeout of the PTP.
Here is the critical part of it:
(1.2 one, this is correct)

[Mario.local:05548]  Map for job: 1 Generated by mapping mode: byslot

  Starting vpid: 0 Vpid range: 4 Num app_contexts: 1

...

...

(1.4 one)

[Mario.local:05542]  Map for job: 0 Generated by mapping mode: byslot

  Starting vpid: 0 Vpid range: 1 Num app_contexts: 1

...

...

Seems the 1.4 mapps the wrong job, Im not sure to what is it referred to,
but hope it will give you some clues.

On Sun, Apr 18, 2010 at 4:07 AM, Ralph Castain  wrote:

> Just to check what is going on, why don't you remove that message passing
> code and just
>
> printf("Hello MPI World from process %d!", my_rank
>
>  in each process? Much more direct - avoids any ambiguity.
>
> Also, be certain that you compile this program for the specific OMPI
> version you are running it under. OMPI is NOT binary compatible across
> releases - you have to recompile the program for the specific release you
> are going to use.
>
>
> On Apr 17, 2010, at 4:52 PM, Mario Ogrizek wrote:
>
> Ofcourse, its the same program, wasnt recompiled for a week.
>
>
> #include 
> #include 
> #include "mpi.h"
>
> int main(int argc, char* argv[]){
> int  my_rank; /* rank of process */
> int  p;   /* number of processes */
> int source;   /* rank of sender */
> int dest; /* rank of receiver */
> int tag=0;/* tag for messages */
> char message[100];/* storage for message */
> MPI_Status status ;   /* return status for receive */
>
>
> /* start up MPI */
>
>
> MPI_Init(, );
>
> /* find out process rank */
> MPI_Comm_rank(MPI_COMM_WORLD, _rank);
>
>
>
> /* find out number of processes */
> MPI_Comm_size(MPI_COMM_WORLD, );
>
>
> if (my_rank !=0){
> /* create message */
> sprintf(message, "Hello MPI World from process %d!", my_rank);
> dest = 0;
> /* use strlen+1 so that '\0' get transmitted */
> MPI_Send(message, strlen(message)+1, MPI_CHAR,
>dest, tag, MPI_COMM_WORLD);
> }
> else{
> printf("Hello MPI World From process 0: Num processes: %d\n",p);
> for (source = 1; source < p; source++) {
> MPI_Recv(message, 100, MPI_CHAR, source, tag,
>   MPI_COMM_WORLD, );
> printf("%s\n",message);
> }
> }
> /* shut down MPI */
> MPI_Finalize();
>
>
>
> return 0;
> }
>
> I triplechecked:
> v1.2 output
> Hello MPI World From process 0: Num processes: 4
> Hello MPI World from process 1!
> Hello MPI World from process 2!
> Hello MPI World from process 3!
>
> v1.4 output:
>
> Hello MPI World From process 0: Num processes: 1
>
> Hello MPI World From process 0: Num processes: 1
>
> Hello MPI World From process 0: Num processes: 1
>
> Hello MPI World From process 0: Num processes: 1
>
>
>
>
>
>
>
> On Sat, Apr 17, 2010 at 9:13 PM, Ralph Castain  wrote:
>
>>
>> On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:
>>
>> Hahaha, ok then that WAS silly! :D
>> So there is no way to utilize both cores with mpi?
>>
>>
>> We are using both cores - it is just that they are on the same node.
>> Unless told otherwise, the processes will use shared memory for
>> communication.
>>
>>
>> Ah well, I'll correct that.
>>
>> From console, im starting a job like this: mpirun -np 4 Program, where i
>> want to run a Program on 4 processors.
>> I was just stumbled when i got same output 4 times, like there are 4
>> processes ranked 0.
>> While with the old version of mpi (1.2) same execution would give 4
>> processes ranked 0..3.
>>
>>
>> And so you should - if not, then there is something wrong. No way mpirun
>> would start 4 processes ranked 0. How are you printing the rank? Are you
>> sure you are getting it correctly?
>>
>>
>>
>> Hope you see my question.
>>
>> On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain  wrote:
>>
>>>
>>> On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:
>>>
>>> I am new to mpi, so I'm sorry for any silly questions.
>>>
>>> My idea was to try to use dual core machine as two nodes. I have a
>>> limited access to a cluster, so this was just for "testing" purposes.
>>> My default hostfile contains usual comments and this two nodes:
>>>
>>> node0
>>> node1
>>>
>>> I thought that each processor is a node for MPI purpose.
>>>
>>>
>>> I'm afraid not - it is just another processor on that node. So you only
>>> have one node as far as OMPI is concerned.
>>>
>>> Im not sure what do you mean with "mpirun cmd line"?
>>>
>>>
>>> How are you starting your job? The usual way is with "mpirun -n N ...".
>>> That is what we mean by the "mpirun cmd line" - i.e., what command are you
>>> using to start your job?
>>>
>>> It sounds like things are actually working correctly. You might look at
>>> "mpirun -h" for possible options of interest.
>>>
>>>
>>>
>>> Regards,
>>>
>>> Mario
>>>
>>> On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain  wrote:
>>>

 On Apr 16, 2010, at 5:08 PM, 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-17 Thread Ralph Castain

On Apr 17, 2010, at 11:17 AM, Mario Ogrizek wrote:

> Hahaha, ok then that WAS silly! :D
> So there is no way to utilize both cores with mpi?

We are using both cores - it is just that they are on the same node. Unless 
told otherwise, the processes will use shared memory for communication.

> 
> Ah well, I'll correct that.
> 
> From console, im starting a job like this: mpirun -np 4 Program, where i want 
> to run a Program on 4 processors.
> I was just stumbled when i got same output 4 times, like there are 4 
> processes ranked 0.
> While with the old version of mpi (1.2) same execution would give 4 processes 
> ranked 0..3.

And so you should - if not, then there is something wrong. No way mpirun would 
start 4 processes ranked 0. How are you printing the rank? Are you sure you are 
getting it correctly?


> 
> Hope you see my question.
> 
> On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain  wrote:
> 
> On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:
> 
>> I am new to mpi, so I'm sorry for any silly questions.
>> 
>> My idea was to try to use dual core machine as two nodes. I have a limited 
>> access to a cluster, so this was just for "testing" purposes.
>> My default hostfile contains usual comments and this two nodes:
>> 
>>> node0
>>> node1
>> I thought that each processor is a node for MPI purpose.
> 
> I'm afraid not - it is just another processor on that node. So you only have 
> one node as far as OMPI is concerned.
> 
>> Im not sure what do you mean with "mpirun cmd line"?
> 
> How are you starting your job? The usual way is with "mpirun -n N ...". That 
> is what we mean by the "mpirun cmd line" - i.e., what command are you using 
> to start your job?
> 
> It sounds like things are actually working correctly. You might look at 
> "mpirun -h" for possible options of interest.
> 
> 
>> 
>> Regards,
>> 
>> Mario
>> 
>> On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain  wrote:
>> 
>> On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote:
>> 
>>> I checked the default MCA param file, and found it was there that was 
>>> (automatically) specified as a relative path, so i changed it.
>>> So now, it works, altho, still something is not right.
>>> Seems like its creating 4 times only 1 process.
>>> Not sure if it has to do something with my hostfile, it contains:
>>> 
>>> node0
>>> node1
>>> 
>>> I am running this on a simple dualcore machine, so i specified it as a 
>>> localhost with two nodes.
>> 
>> I don't understand this comment - a dual core machine would still be a 
>> single node. Just happens to have two processors in it.
>> 
>> Could you send the contents of your hostfile and your mpirun cmd line?
>> 
>>> 
>>> Regards,
>>> 
>>> Mario
>>> 
>>> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek  
>>> wrote:
>>> I understand, so, its looking for a 
>>> working_dir/usr/local/etc/openmpi-default-hostfile
>>> I managed to run a hello world program from the console, while my wd was 
>>> just "/" and it worked, altho strangely...
>>> example for 4 procs:
>>> 
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> 
>>> So, you are saying i allways have to be in "/" to run mpi programs, or 
>>> there is a way for mpi to search absolute path?
>>> It seems pretty inconvinient this way.
>>> I think v 1.2 didnt have this limitation.
>>> 
>>> Does this have to do anything with LD_LIBRARY_PATH?
>>> 
>>> Regards,
>>> 
>>> Mario
>>> 
>>> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain  wrote:
>>> How did you specify it? Command line? Default MCA param file?
>>> 
>>> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
>>> 
 Any idea how to solve this?
 
 On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov 
  wrote:
 Hello.
 It looks that you hostfile path should
 be /usr/local/etc/openmpi-default-hostfile not
 usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
 second path.
 
 В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
 > Well, im not sure why should i name it /openmpi-default-hostfile
 > Especially, because mpirun v1.2 executes without any errors.
 > But, i made a copy named /openmpi-default-hostfile, and still, the
 > same result.
 >
 > This is the whole error message for a simple hello world program:
 >
 >
 > Open RTE was unable to open the hostfile:
 > usr/local/etc/openmpi-default-hostfile
 > Check to make sure the path and filename are correct.
 > --
 > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
 > base/ras_base_allocate.c at line 186
 > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
 > 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-17 Thread Mario Ogrizek
Hahaha, ok then that WAS silly! :D
So there is no way to utilize both cores with mpi?

Ah well, I'll correct that.

>From console, im starting a job like this: mpirun -np 4 Program, where i
want to run a Program on 4 processors.
I was just stumbled when i got same output 4 times, like there are 4
processes ranked 0.
While with the old version of mpi (1.2) same execution would give 4
processes ranked 0..3.

Hope you see my question.

On Sat, Apr 17, 2010 at 6:29 PM, Ralph Castain  wrote:

>
> On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:
>
> I am new to mpi, so I'm sorry for any silly questions.
>
> My idea was to try to use dual core machine as two nodes. I have a limited
> access to a cluster, so this was just for "testing" purposes.
> My default hostfile contains usual comments and this two nodes:
>
> node0
> node1
>
> I thought that each processor is a node for MPI purpose.
>
>
> I'm afraid not - it is just another processor on that node. So you only
> have one node as far as OMPI is concerned.
>
> Im not sure what do you mean with "mpirun cmd line"?
>
>
> How are you starting your job? The usual way is with "mpirun -n N ...".
> That is what we mean by the "mpirun cmd line" - i.e., what command are you
> using to start your job?
>
> It sounds like things are actually working correctly. You might look at
> "mpirun -h" for possible options of interest.
>
>
>
> Regards,
>
> Mario
>
> On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain  wrote:
>
>>
>> On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote:
>>
>> I checked the default MCA param file, and found it was there that was
>> (automatically) specified as a relative path, so i changed it.
>> So now, it works, altho, still something is not right.
>> Seems like its creating 4 times only 1 process.
>> Not sure if it has to do something with my hostfile, it contains:
>>
>> node0
>> node1
>>
>> I am running this on a simple dualcore machine, so i specified it as a
>> localhost with two nodes.
>>
>>
>> I don't understand this comment - a dual core machine would still be a
>> single node. Just happens to have two processors in it.
>>
>> Could you send the contents of your hostfile and your mpirun cmd line?
>>
>>
>> Regards,
>>
>> Mario
>>
>> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek > > wrote:
>>
>>> I understand, so, its looking for a
>>> working_dir/usr/local/etc/openmpi-default-hostfile
>>> I managed to run a hello world program from the console, while my wd was
>>> just "/" and it worked, altho strangely...
>>> example for 4 procs:
>>>
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>> Hello MPI World From process 0: Num processes: 1
>>>
>>> So, you are saying i allways have to be in "/" to run mpi programs, or
>>> there is a way for mpi to search absolute path?
>>> It seems pretty inconvinient this way.
>>> I think v 1.2 didnt have this limitation.
>>>
>>> Does this have to do anything with LD_LIBRARY_PATH?
>>>
>>> Regards,
>>>
>>> Mario
>>>
>>> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain  wrote:
>>>
 How did you specify it? Command line? Default MCA param file?

 On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:

 Any idea how to solve this?

 On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov <
 timur.magome...@developonbox.ru> wrote:

> Hello.
> It looks that you hostfile path should
> be /usr/local/etc/openmpi-default-hostfile not
> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
> second path.
>
> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
> > Well, im not sure why should i name it /openmpi-default-hostfile
> > Especially, because mpirun v1.2 executes without any errors.
> > But, i made a copy named /openmpi-default-hostfile, and still, the
> > same result.
> >
> > This is the whole error message for a simple hello world program:
> >
> >
> > Open RTE was unable to open the hostfile:
> > usr/local/etc/openmpi-default-hostfile
> > Check to make sure the path and filename are correct.
> >
> --
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > base/ras_base_allocate.c at line 186
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > base/plm_base_launch_support.c at line 72
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > plm_rsh_module.c at line 990
> >
> --
> > A daemon (pid unknown) died unexpectedly on signal 1  while
> attempting
> > to
> > launch so we are aborting.
> >
> >
> > There may be more information reported by the environment 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-17 Thread Ralph Castain

On Apr 17, 2010, at 1:16 AM, Mario Ogrizek wrote:

> I am new to mpi, so I'm sorry for any silly questions.
> 
> My idea was to try to use dual core machine as two nodes. I have a limited 
> access to a cluster, so this was just for "testing" purposes.
> My default hostfile contains usual comments and this two nodes:
> 
>> node0
>> node1
> I thought that each processor is a node for MPI purpose.

I'm afraid not - it is just another processor on that node. So you only have 
one node as far as OMPI is concerned.

> Im not sure what do you mean with "mpirun cmd line"?

How are you starting your job? The usual way is with "mpirun -n N ...". That is 
what we mean by the "mpirun cmd line" - i.e., what command are you using to 
start your job?

It sounds like things are actually working correctly. You might look at "mpirun 
-h" for possible options of interest.


> 
> Regards,
> 
> Mario
> 
> On Sat, Apr 17, 2010 at 1:54 AM, Ralph Castain  wrote:
> 
> On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote:
> 
>> I checked the default MCA param file, and found it was there that was 
>> (automatically) specified as a relative path, so i changed it.
>> So now, it works, altho, still something is not right.
>> Seems like its creating 4 times only 1 process.
>> Not sure if it has to do something with my hostfile, it contains:
>> 
>> node0
>> node1
>> 
>> I am running this on a simple dualcore machine, so i specified it as a 
>> localhost with two nodes.
> 
> I don't understand this comment - a dual core machine would still be a single 
> node. Just happens to have two processors in it.
> 
> Could you send the contents of your hostfile and your mpirun cmd line?
> 
>> 
>> Regards,
>> 
>> Mario
>> 
>> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek  
>> wrote:
>> I understand, so, its looking for a 
>> working_dir/usr/local/etc/openmpi-default-hostfile
>> I managed to run a hello world program from the console, while my wd was 
>> just "/" and it worked, altho strangely...
>> example for 4 procs:
>> 
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> Hello MPI World From process 0: Num processes: 1
>> 
>> So, you are saying i allways have to be in "/" to run mpi programs, or there 
>> is a way for mpi to search absolute path?
>> It seems pretty inconvinient this way.
>> I think v 1.2 didnt have this limitation.
>> 
>> Does this have to do anything with LD_LIBRARY_PATH?
>> 
>> Regards,
>> 
>> Mario
>> 
>> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain  wrote:
>> How did you specify it? Command line? Default MCA param file?
>> 
>> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
>> 
>>> Any idea how to solve this?
>>> 
>>> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov 
>>>  wrote:
>>> Hello.
>>> It looks that you hostfile path should
>>> be /usr/local/etc/openmpi-default-hostfile not
>>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
>>> second path.
>>> 
>>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
>>> > Well, im not sure why should i name it /openmpi-default-hostfile
>>> > Especially, because mpirun v1.2 executes without any errors.
>>> > But, i made a copy named /openmpi-default-hostfile, and still, the
>>> > same result.
>>> >
>>> > This is the whole error message for a simple hello world program:
>>> >
>>> >
>>> > Open RTE was unable to open the hostfile:
>>> > usr/local/etc/openmpi-default-hostfile
>>> > Check to make sure the path and filename are correct.
>>> > --
>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>> > base/ras_base_allocate.c at line 186
>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>> > base/plm_base_launch_support.c at line 72
>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>> > plm_rsh_module.c at line 990
>>> > --
>>> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
>>> > to
>>> > launch so we are aborting.
>>> >
>>> >
>>> > There may be more information reported by the environment (see above).
>>> >
>>> >
>>> > This may be because the daemon was unable to find all the needed
>>> > shared
>>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>>> > the
>>> > location of the shared libraries on the remote nodes and this will
>>> > automatically be forwarded to the remote nodes.
>>> > --
>>> > --
>>> > mpirun noticed that the job aborted, but has no info as to the process
>>> > that caused that situation.
>>> > 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Ralph Castain

On Apr 16, 2010, at 5:08 PM, Mario Ogrizek wrote:

> I checked the default MCA param file, and found it was there that was 
> (automatically) specified as a relative path, so i changed it.
> So now, it works, altho, still something is not right.
> Seems like its creating 4 times only 1 process.
> Not sure if it has to do something with my hostfile, it contains:
> 
> node0
> node1
> 
> I am running this on a simple dualcore machine, so i specified it as a 
> localhost with two nodes.

I don't understand this comment - a dual core machine would still be a single 
node. Just happens to have two processors in it.

Could you send the contents of your hostfile and your mpirun cmd line?

> 
> Regards,
> 
> Mario
> 
> On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek  
> wrote:
> I understand, so, its looking for a 
> working_dir/usr/local/etc/openmpi-default-hostfile
> I managed to run a hello world program from the console, while my wd was just 
> "/" and it worked, altho strangely...
> example for 4 procs:
> 
> Hello MPI World From process 0: Num processes: 1
> Hello MPI World From process 0: Num processes: 1
> Hello MPI World From process 0: Num processes: 1
> Hello MPI World From process 0: Num processes: 1
> 
> So, you are saying i allways have to be in "/" to run mpi programs, or there 
> is a way for mpi to search absolute path?
> It seems pretty inconvinient this way.
> I think v 1.2 didnt have this limitation.
> 
> Does this have to do anything with LD_LIBRARY_PATH?
> 
> Regards,
> 
> Mario
> 
> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain  wrote:
> How did you specify it? Command line? Default MCA param file?
> 
> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
> 
>> Any idea how to solve this?
>> 
>> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov 
>>  wrote:
>> Hello.
>> It looks that you hostfile path should
>> be /usr/local/etc/openmpi-default-hostfile not
>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
>> second path.
>> 
>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
>> > Well, im not sure why should i name it /openmpi-default-hostfile
>> > Especially, because mpirun v1.2 executes without any errors.
>> > But, i made a copy named /openmpi-default-hostfile, and still, the
>> > same result.
>> >
>> > This is the whole error message for a simple hello world program:
>> >
>> >
>> > Open RTE was unable to open the hostfile:
>> > usr/local/etc/openmpi-default-hostfile
>> > Check to make sure the path and filename are correct.
>> > --
>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>> > base/ras_base_allocate.c at line 186
>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>> > base/plm_base_launch_support.c at line 72
>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>> > plm_rsh_module.c at line 990
>> > --
>> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
>> > to
>> > launch so we are aborting.
>> >
>> >
>> > There may be more information reported by the environment (see above).
>> >
>> >
>> > This may be because the daemon was unable to find all the needed
>> > shared
>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>> > the
>> > location of the shared libraries on the remote nodes and this will
>> > automatically be forwarded to the remote nodes.
>> > --
>> > --
>> > mpirun noticed that the job aborted, but has no info as to the process
>> > that caused that situation.
>> > --
>> > mpirun: clean termination accomplished
>> >
>> >
>> >
>> >
>> > ps. PTP is a parallel tools platform plugin for eclipse
>> >
>> >
>> > Regards,
>> >
>> >
>> > Mario
>> >
>> > ___
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> --
>> Kind regards,
>> Timur Magomedov
>> Senior C++ Developer
>> DevelopOnBox LLC / Zodiac Interactive
>> http://www.zodiac.tv/
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> 

Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Mario Ogrizek
I checked the default MCA param file, and found it was there that was
(automatically) specified as a relative path, so i changed it.
So now, it works, altho, still something is not right.
Seems like its creating 4 times only 1 process.
Not sure if it has to do something with my hostfile, it contains:

node0
node1

I am running this on a simple dualcore machine, so i specified it as a
localhost with two nodes.

Regards,

Mario

On Sat, Apr 17, 2010 at 12:52 AM, Mario Ogrizek wrote:

> I understand, so, its looking for a
> working_dir/usr/local/etc/openmpi-default-hostfile
> I managed to run a hello world program from the console, while my wd was
> just "/" and it worked, altho strangely...
> example for 4 procs:
>
> Hello MPI World From process 0: Num processes: 1
> Hello MPI World From process 0: Num processes: 1
> Hello MPI World From process 0: Num processes: 1
> Hello MPI World From process 0: Num processes: 1
>
> So, you are saying i allways have to be in "/" to run mpi programs, or
> there is a way for mpi to search absolute path?
> It seems pretty inconvinient this way.
> I think v 1.2 didnt have this limitation.
>
> Does this have to do anything with LD_LIBRARY_PATH?
>
> Regards,
>
> Mario
>
> On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain  wrote:
>
>> How did you specify it? Command line? Default MCA param file?
>>
>> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
>>
>> Any idea how to solve this?
>>
>> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov <
>> timur.magome...@developonbox.ru> wrote:
>>
>>> Hello.
>>> It looks that you hostfile path should
>>> be /usr/local/etc/openmpi-default-hostfile not
>>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
>>> second path.
>>>
>>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
>>> > Well, im not sure why should i name it /openmpi-default-hostfile
>>> > Especially, because mpirun v1.2 executes without any errors.
>>> > But, i made a copy named /openmpi-default-hostfile, and still, the
>>> > same result.
>>> >
>>> > This is the whole error message for a simple hello world program:
>>> >
>>> >
>>> > Open RTE was unable to open the hostfile:
>>> > usr/local/etc/openmpi-default-hostfile
>>> > Check to make sure the path and filename are correct.
>>> >
>>> --
>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>> > base/ras_base_allocate.c at line 186
>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>> > base/plm_base_launch_support.c at line 72
>>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>>> > plm_rsh_module.c at line 990
>>> >
>>> --
>>> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
>>> > to
>>> > launch so we are aborting.
>>> >
>>> >
>>> > There may be more information reported by the environment (see above).
>>> >
>>> >
>>> > This may be because the daemon was unable to find all the needed
>>> > shared
>>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>>> > the
>>> > location of the shared libraries on the remote nodes and this will
>>> > automatically be forwarded to the remote nodes.
>>> >
>>> --
>>> >
>>> --
>>> > mpirun noticed that the job aborted, but has no info as to the process
>>> > that caused that situation.
>>> >
>>> --
>>> > mpirun: clean termination accomplished
>>> >
>>> >
>>> >
>>> >
>>> > ps. PTP is a parallel tools platform plugin for eclipse
>>> >
>>> >
>>> > Regards,
>>> >
>>> >
>>> > Mario
>>> >
>>> > ___
>>> > users mailing list
>>> > us...@open-mpi.org
>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>>
>>> --
>>> Kind regards,
>>> Timur Magomedov
>>> Senior C++ Developer
>>> DevelopOnBox LLC / Zodiac Interactive
>>> http://www.zodiac.tv/
>>>
>>> ___
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>>
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>
>


Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Mario Ogrizek
I understand, so, its looking for a
working_dir/usr/local/etc/openmpi-default-hostfile
I managed to run a hello world program from the console, while my wd was
just "/" and it worked, altho strangely...
example for 4 procs:

Hello MPI World From process 0: Num processes: 1
Hello MPI World From process 0: Num processes: 1
Hello MPI World From process 0: Num processes: 1
Hello MPI World From process 0: Num processes: 1

So, you are saying i allways have to be in "/" to run mpi programs, or there
is a way for mpi to search absolute path?
It seems pretty inconvinient this way.
I think v 1.2 didnt have this limitation.

Does this have to do anything with LD_LIBRARY_PATH?

Regards,

Mario

On Fri, Apr 16, 2010 at 7:46 PM, Ralph Castain  wrote:

> How did you specify it? Command line? Default MCA param file?
>
> On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:
>
> Any idea how to solve this?
>
> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov <
> timur.magome...@developonbox.ru> wrote:
>
>> Hello.
>> It looks that you hostfile path should
>> be /usr/local/etc/openmpi-default-hostfile not
>> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
>> second path.
>>
>> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
>> > Well, im not sure why should i name it /openmpi-default-hostfile
>> > Especially, because mpirun v1.2 executes without any errors.
>> > But, i made a copy named /openmpi-default-hostfile, and still, the
>> > same result.
>> >
>> > This is the whole error message for a simple hello world program:
>> >
>> >
>> > Open RTE was unable to open the hostfile:
>> > usr/local/etc/openmpi-default-hostfile
>> > Check to make sure the path and filename are correct.
>> >
>> --
>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>> > base/ras_base_allocate.c at line 186
>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>> > base/plm_base_launch_support.c at line 72
>> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
>> > plm_rsh_module.c at line 990
>> >
>> --
>> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
>> > to
>> > launch so we are aborting.
>> >
>> >
>> > There may be more information reported by the environment (see above).
>> >
>> >
>> > This may be because the daemon was unable to find all the needed
>> > shared
>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
>> > the
>> > location of the shared libraries on the remote nodes and this will
>> > automatically be forwarded to the remote nodes.
>> >
>> --
>> >
>> --
>> > mpirun noticed that the job aborted, but has no info as to the process
>> > that caused that situation.
>> >
>> --
>> > mpirun: clean termination accomplished
>> >
>> >
>> >
>> >
>> > ps. PTP is a parallel tools platform plugin for eclipse
>> >
>> >
>> > Regards,
>> >
>> >
>> > Mario
>> >
>> > ___
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>> --
>> Kind regards,
>> Timur Magomedov
>> Senior C++ Developer
>> DevelopOnBox LLC / Zodiac Interactive
>> http://www.zodiac.tv/
>>
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Ralph Castain
How did you specify it? Command line? Default MCA param file?

On Apr 16, 2010, at 11:44 AM, Mario Ogrizek wrote:

> Any idea how to solve this?
> 
> On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov 
>  wrote:
> Hello.
> It looks that you hostfile path should
> be /usr/local/etc/openmpi-default-hostfile not
> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
> second path.
> 
> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
> > Well, im not sure why should i name it /openmpi-default-hostfile
> > Especially, because mpirun v1.2 executes without any errors.
> > But, i made a copy named /openmpi-default-hostfile, and still, the
> > same result.
> >
> > This is the whole error message for a simple hello world program:
> >
> >
> > Open RTE was unable to open the hostfile:
> > usr/local/etc/openmpi-default-hostfile
> > Check to make sure the path and filename are correct.
> > --
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > base/ras_base_allocate.c at line 186
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > base/plm_base_launch_support.c at line 72
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > plm_rsh_module.c at line 990
> > --
> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
> > to
> > launch so we are aborting.
> >
> >
> > There may be more information reported by the environment (see above).
> >
> >
> > This may be because the daemon was unable to find all the needed
> > shared
> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
> > the
> > location of the shared libraries on the remote nodes and this will
> > automatically be forwarded to the remote nodes.
> > --
> > --
> > mpirun noticed that the job aborted, but has no info as to the process
> > that caused that situation.
> > --
> > mpirun: clean termination accomplished
> >
> >
> >
> >
> > ps. PTP is a parallel tools platform plugin for eclipse
> >
> >
> > Regards,
> >
> >
> > Mario
> >
> > ___
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> --
> Kind regards,
> Timur Magomedov
> Senior C++ Developer
> DevelopOnBox LLC / Zodiac Interactive
> http://www.zodiac.tv/
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Ralph Castain
The problem is that you gave us a relative path - is that where the file is 
located?

The system is looking for usr/local/etc/openmpi-default-hostfile relative to 
your current working directory. If you want us to look in /usr/local/etc, then 
you have to give us that absolute path.

We don't care what the filename is - we do care that we have a correct path to 
it.

On Apr 16, 2010, at 11:10 AM, Mario Ogrizek wrote:

> Well, im not sure why should i name it /openmpi-default-hostfile
> Especially, because mpirun v1.2 executes without any errors.
> But, i made a copy named /openmpi-default-hostfile, and still, the same 
> result.
> 
> This is the whole error message for a simple hello world program:
> 
> Open RTE was unable to open the hostfile:
> usr/local/etc/openmpi-default-hostfile
> Check to make sure the path and filename are correct.
> --
> [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file 
> base/ras_base_allocate.c at line 186
> [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file 
> base/plm_base_launch_support.c at line 72
> [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file 
> plm_rsh_module.c at line 990
> --
> A daemon (pid unknown) died unexpectedly on signal 1  while attempting to
> launch so we are aborting.
> 
> There may be more information reported by the environment (see above).
> 
> This may be because the daemon was unable to find all the needed shared
> libraries on the remote node. You may set your LD_LIBRARY_PATH to have the
> location of the shared libraries on the remote nodes and this will
> automatically be forwarded to the remote nodes.
> --
> --
> mpirun noticed that the job aborted, but has no info as to the process
> that caused that situation.
> --
> mpirun: clean termination accomplished
> 
> 
> ps. PTP is a parallel tools platform plugin for eclipse 
> 
> Regards,
> 
> Mario
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Mario Ogrizek
Any idea how to solve this?

On Fri, Apr 16, 2010 at 7:40 PM, Timur Magomedov <
timur.magome...@developonbox.ru> wrote:

> Hello.
> It looks that you hostfile path should
> be /usr/local/etc/openmpi-default-hostfile not
> usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
> second path.
>
> В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
> > Well, im not sure why should i name it /openmpi-default-hostfile
> > Especially, because mpirun v1.2 executes without any errors.
> > But, i made a copy named /openmpi-default-hostfile, and still, the
> > same result.
> >
> > This is the whole error message for a simple hello world program:
> >
> >
> > Open RTE was unable to open the hostfile:
> > usr/local/etc/openmpi-default-hostfile
> > Check to make sure the path and filename are correct.
> >
> --
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > base/ras_base_allocate.c at line 186
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > base/plm_base_launch_support.c at line 72
> > [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> > plm_rsh_module.c at line 990
> >
> --
> > A daemon (pid unknown) died unexpectedly on signal 1  while attempting
> > to
> > launch so we are aborting.
> >
> >
> > There may be more information reported by the environment (see above).
> >
> >
> > This may be because the daemon was unable to find all the needed
> > shared
> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
> > the
> > location of the shared libraries on the remote nodes and this will
> > automatically be forwarded to the remote nodes.
> >
> --
> >
> --
> > mpirun noticed that the job aborted, but has no info as to the process
> > that caused that situation.
> >
> --
> > mpirun: clean termination accomplished
> >
> >
> >
> >
> > ps. PTP is a parallel tools platform plugin for eclipse
> >
> >
> > Regards,
> >
> >
> > Mario
> >
> > ___
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Kind regards,
> Timur Magomedov
> Senior C++ Developer
> DevelopOnBox LLC / Zodiac Interactive
> http://www.zodiac.tv/
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


Re: [OMPI users] Fwd: Open MPI v1.4 cant find default hostfile

2010-04-16 Thread Timur Magomedov
Hello.
It looks that you hostfile path should
be /usr/local/etc/openmpi-default-hostfile not
usr/local/etc/openmpi-default-hostfile but somehow Open MPI gets the
second path.

В Птн, 16/04/2010 в 19:10 +0200, Mario Ogrizek пишет:
> Well, im not sure why should i name it /openmpi-default-hostfile
> Especially, because mpirun v1.2 executes without any errors.
> But, i made a copy named /openmpi-default-hostfile, and still, the
> same result.
> 
> This is the whole error message for a simple hello world program: 
> 
> 
> Open RTE was unable to open the hostfile:
> usr/local/etc/openmpi-default-hostfile
> Check to make sure the path and filename are correct.
> --
> [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> base/ras_base_allocate.c at line 186
> [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> base/plm_base_launch_support.c at line 72
> [Mario.local:04300] [[114,0],0] ORTE_ERROR_LOG: Not found in file
> plm_rsh_module.c at line 990
> --
> A daemon (pid unknown) died unexpectedly on signal 1  while attempting
> to
> launch so we are aborting.
> 
> 
> There may be more information reported by the environment (see above).
> 
> 
> This may be because the daemon was unable to find all the needed
> shared
> libraries on the remote node. You may set your LD_LIBRARY_PATH to have
> the
> location of the shared libraries on the remote nodes and this will
> automatically be forwarded to the remote nodes.
> --
> --
> mpirun noticed that the job aborted, but has no info as to the process
> that caused that situation.
> --
> mpirun: clean termination accomplished
> 
> 
> 
> 
> ps. PTP is a parallel tools platform plugin for eclipse 
> 
> 
> Regards,
> 
> 
> Mario
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Kind regards,
Timur Magomedov
Senior C++ Developer
DevelopOnBox LLC / Zodiac Interactive
http://www.zodiac.tv/