can you update me with the mapping or the way to get it from the OS on the
Cell.

thanks

On Thu, Oct 23, 2008 at 8:08 PM, Mi Yan <mi...@us.ibm.com> wrote:

> Lenny,
>
> Thanks.
> I asked the Cell/BE Linux Kernel developer to get the CPU mapping :) The
> mapping is fixed in current kernel.
>
> Mi
> [image: Inactive hide details for "Lenny Verkhovsky"
> <lenny.verkhov...@gmail.com>]"Lenny Verkhovsky" <
> lenny.verkhov...@gmail.com>
>
>
>
>     *"Lenny Verkhovsky" <lenny.verkhov...@gmail.com>*
>             Sent by: users-boun...@open-mpi.org
>
>             10/23/2008 01:52 PM Please respond to
>             Open MPI Users <us...@open-mpi.org>
>
>
> To
>
> "Open MPI Users" <us...@open-mpi.org>
> cc
>
>
> Subject
>
> Re: [OMPI users] Working with a CellBlade cluster
> According to *
> https://svn.open-mpi.org/trac/ompi/milestone/Open%20MPI%201.3*<https://svn.open-mpi.org/trac/ompi/milestone/Open%20MPI%201.3>very
>  soon,
> but you can download trunk version 
> *http://www.open-mpi.org/svn/*<http://www.open-mpi.org/svn/>and check if it 
> works for you.
>
> how can you check mapping CPUs by OS , my cat /proc/cpuinfo shows very
> little info
> # cat /proc/cpuinfo
> processor : 0
> cpu : Cell Broadband Engine, altivec supported
> clock : 3200.000000MHz
> revision : 48.0 (pvr 0070 3000)
> processor : 1
> cpu : Cell Broadband Engine, altivec supported
> clock : 3200.000000MHz
> revision : 48.0 (pvr 0070 3000)
> processor : 2
> cpu : Cell Broadband Engine, altivec supported
> clock : 3200.000000MHz
> revision : 48.0 (pvr 0070 3000)
> processor : 3
> cpu : Cell Broadband Engine, altivec supported
> clock : 3200.000000MHz
> revision : 48.0 (pvr 0070 3000)
> timebase : 26666666
> platform : Cell
> machine : CHRP IBM,0793-1RZ
>
>
>
> On Thu, Oct 23, 2008 at 3:00 PM, Mi Yan <*mi...@us.ibm.com*<mi...@us.ibm.com>>
> wrote:
>
>    Hi, Lenny,
>
>    So rank file map will be supported in OpenMPI 1.3? I'm using
>    OpenMPI1.2.6 and did not find parameter "rmaps_rank_file_".
>    Do you have idea when OpenMPI 1.3 will be available? OpenMPI 1.3 has
>    quite a few features I'm looking for.
>
>    Thanks,
>
>    Mi
>    [image: Inactive hide details for "Lenny Verkhovsky"
>    <lenny.verkhov...@gmail.com>]"Lenny Verkhovsky" <*
>    lenny.verkhov...@gmail.com* <lenny.verkhov...@gmail.com>>
>
>
>
>
>       *"Lenny Verkhovsky" 
> <**lenny.verkhov...@gmail.com*<lenny.verkhov...@gmail.com>
>                         *>*
>                         Sent by: 
> *users-boun...@open-mpi.org*<users-boun...@open-mpi.org>
>
>                         10/23/2008 05:48 AM
>
> Please respond to
> Open MPI Users <*us...@open-mpi.org* <us...@open-mpi.org>>
>  To
>
> "Open MPI Users" <*us...@open-mpi.org* <us...@open-mpi.org>>cc
> Subject
>
> Re: [OMPI users] Working with a CellBlade cluster
>
>
>    Hi,
>
>
>    If I understand you correctly the most suitable way to do it is by
>    paffinity that we have in Open MPI 1.3 and the trank.
>    how ever usually OS is distributing processes evenly between sockets by
>    it self.
>
>    There still no formal FAQ due to a multiple reasons but you can read
>    how to use it in the attached scratch ( there were few name changings of 
> the
>    params, so check with ompi_info )
>
>    shared memory is used between processes that share same machine, and
>    openib is used between different machines ( hostnames ), no special mca
>    params are needed.
>
>    Best Regards
>    Lenny,
>
>
>     On Sun, Oct 19, 2008 at 10:32 AM, Gilbert Grosdidier <*
>    gro...@mail.cern.ch* <gro...@mail.cern.ch>> wrote:
>       Working with a CellBlade cluster (QS22), the requirement is to have
>          one
>          instance of the executable running on each socket of the blade
>          (there are 2
>          sockets). The application is of the 'domain decomposition' type,
>          and each
>          instance is required to often send/receive data with both the
>          remote blades and
>          the neighbor socket.
>
>          Question is : which specification must be used for the mca btl
>          component
>          to force 1) shmem type messages when communicating with this
>          neighbor socket,
>          while 2) using openib to communicate with the remote blades ?
>          Is '-mca btl sm,openib,self' suitable for this ?
>
>          Also, which debug flags could be used to crosscheck that the
>          messages are
>          _actually_ going thru the right channel for a given channel,
>          please ?
>
>          We are currently using OpenMPI 1.2.5 shipped with RHEL5.2
>          (ppc64).
>          Which version do you think is currently the most optimised for
>          these
>          processors and problem type ? Should we go towards OpenMPI 1.2.8
>          instead ?
>          Or even try some OpenMPI 1.3 nightly build ?
>
>          Thanks in advance for your help, Gilbert.
>
>          _______________________________________________
>          users mailing list*
>          **us...@open-mpi.org* <us...@open-mpi.org>*
>          
> **http://www.open-mpi.org/mailman/listinfo.cgi/users*<http://www.open-mpi.org/mailman/listinfo.cgi/users>
>        *(See attached file: RANKS_FAQ.doc)*
>    _______________________________________________
>
>    users mailing list*
>    **us...@open-mpi.org* <us...@open-mpi.org>
>    
> *http://www.open-mpi.org/mailman/listinfo.cgi/users*<http://www.open-mpi.org/mailman/listinfo.cgi/users>
>
>    _______________________________________________
>    users mailing list*
>    **us...@open-mpi.org* <us...@open-mpi.org>*
>    
> **http://www.open-mpi.org/mailman/listinfo.cgi/users*<http://www.open-mpi.org/mailman/listinfo.cgi/users>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to