Not that I know of - I don't think we currently have any way for you to
specify a location for individual nodes.
Is there some reason why you installed it this way?
On Fri, Aug 7, 2009 at 11:27 AM, Kenneth Yoshimoto wrote:
>
> Hello,
>
> I have three sets of nodes, each with openmpi installed
Gus Correa wrote:
Terry Frankcombe wrote:
There's been quite some discussion here lately about the effect of OMPI
tuning parameters, particularly w.r.t. collectives.
Is there some tool to probe performance on any given network/collection
of nodes to aid optimisation of these parameters?
(I'm t
I have several Sun x4100 with Infiniband which appear to be running at
400MB/sec instead of 800MB/sec. It a freshly reformatted cluster
converting from solaris to linux. We also reset the bios settings
with "load optimal defaults". Does anyone know which bios setting i
changed to dump the BW?
x4
I'm not sure that there is a general "best set" of parameters, given the
dependence of that set on comms patterns, etc...
Still, this *is* a classic parameter sweep and optimization problem
(unlike ATLAS), with a small number of parameters, and is the sort of
thing one should be able to hook u
Terry Frankcombe wrote:
There's been quite some discussion here lately about the effect of OMPI
tuning parameters, particularly w.r.t. collectives.
Is there some tool to probe performance on any given network/collection
of nodes to aid optimisation of these parameters?
(I'm thinking something a
Hello,
I have three sets of nodes, each with openmpi installed in
a different location. I am getting an error related to orted:
/users/kenneth/info/openmpi/install/bin/orted: Command not found.
I think it's looking for orted in the wrong place on some of
the nodes. Is there an easy way to
Hi Craig, Terry, Neeraj, list
Craig: A fellow here runs WRF.
I grep the code and there is plenty of collectives there:
MPI_[All]Gather[v], MPI_[All]Reduce, etc.
Domain decomposition code like WRF, MITgcm, and other atmosphere
and ocean codes has point-to-point communication to exchange
subdomain
Rolf Vandevaart wrote:
This message is telling you that you have run out of file descriptors.
I am surprised that the -mca parameter setting did not fix the problem.
Can you run limit or ulimit on your shell and send the information? I
typically set my limit to 65536 assuming the system allows
This message is telling you that you have run out of file descriptors.
I am surprised that the -mca parameter setting did not fix the problem.
Can you run limit or ulimit on your shell and send the information? I
typically set my limit to 65536 assuming the system allows it.
burl-16 58 =>limit
Hello all,
I'm using hpc8.2 :
Lidia-jobic% ompi_info
Displaying Open MPI information for 32-bit ...
Package: ClusterTools 8.2
Open MPI: 1.3.3r21324-ct8.2-b09j-r40
[...]
And i've got a X4600 machine (8*4 cores).
When i'm trying to run a 32 processor jobs, i've got
To use tuned collectives, do all I have to do is add --mca coll tuned?
I am trying to run with:
# mpirun -np 8 --mca coll tuned --mca orte_base_help_aggregate 0 ./wrf.exe
But all the processes fail with the folling message:
--
Terry Dontje wrote:
Craig,
Did your affinity script bind the processes per socket or linearly to
cores. If the former you'll want to look at using rankfiles and place
the ranks based on sockets. TWe have found this especially useful if
you are not running fully subscribed on your machines.
Date: Fri, 07 Aug 2009 07:12:45 -0600
From: Craig Tierney
Subject: Re: [OMPI users] Performance question about OpenMPI and
MVAPICH2 on IB
To: Open MPI Users
Message-ID: <4a7c284d.3040...@noaa.gov>
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
Terry Dontje wrote:
>
I'm way behind on my mail; apologies for the delay in replying.
Did you figure this out?
As a pure guess, it sounds like you have a heterogeneous setup --
nodes have different distros and/or versions. As such, your glibc's
may be different, etc. In such situations, it is definitely
recom
nee...@crlindia.com wrote:
Hi Craig,
How was the nodefile selected for execution? Whether it was
provided by scheduler say LSF/SGE/PBS or you manually gave it?
With WRF, we observed giving sequential nodes (Blades which are in the
same order as in enclosure) gave us some performance be
Terry Dontje wrote:
Craig,
Did your affinity script bind the processes per socket or linearly to
cores. If the former you'll want to look at using rankfiles and place
the ranks based on sockets. TWe have found this especially useful if
you are not running fully subscribed on your machines.
Hi Terry,
I had tested mostly MPI_Bcast, MPI_Reduce, MPI_Gather kind of MPI
with openmpi-1.3 and hierarchical option enabled.In all these, i found
results slower than regular tuned collectives.
We have HP Blade with intel clovertown processor(two quad core)
connected with DDR i
Hi Neeraj,
Were there specific collectives that were slower? Also what kind of cluster
were you running on? How many nodes and cores per node?
thanks,
--td
Message: 3
Date: Fri, 7 Aug 2009 16:51:05 +0530
From: nee...@crlindia.com
Subject: Re: [OMPI users] Performance question about OpenMPI
Hi Craig,
How was the nodefile selected for execution? Whether it was
provided by scheduler say LSF/SGE/PBS or you manually gave it?
With WRF, we observed giving sequential nodes (Blades which are in the
same order as in enclosure) gave us some performance benefit.
Regards
Neeraj Chour
Hi Terry,
I feel hierarchical collectives are slower compare to tuned one. I
had done some benchmark in the past specific to collectives, and this is
what i feel based on my observation.
Regards
Neeraj Chourasia (MTS)
Computational Research Laboratories Ltd.
(A wholly Owned Subsidiary
Craig,
Did your affinity script bind the processes per socket or linearly to cores.
If the former you'll want to look at using rankfiles and place the ranks based
on sockets. TWe have found this especially useful if you are not running fully
subscribed on your machines.
Also, if you think t
There's been quite some discussion here lately about the effect of OMPI
tuning parameters, particularly w.r.t. collectives.
Is there some tool to probe performance on any given network/collection
of nodes to aid optimisation of these parameters?
(I'm thinking something along the philosophy of ATL
Hi Craig,
WRF has pattern of talking to nearest neighbours like p+1, p-1,
p+3 and p-3, where p is the particular process. But in addition to that,
it also uses collective calls like MPI_Bcast, MPI_AlltoAllv,
MPI_Allgather, MPI_Gather, MPI_Gatherv, MPI_Scatterv.
Apparently openm
may be because of insufficient storage space...???( I mean hard
disk space.)
On Thu, Aug 6, 2009 at 11:23 PM, Jeff Squyres wrote:
> Any chance you could re-try the experiment with Open MPI 1.3.3?
>
>
> On Aug 4, 2009, at 11:10 AM, Hoelzlwimmer Andreas - S0810595005 wrote:
>
> Hello,
24 matches
Mail list logo