It seems to be working fine for me:

[rhc@bend001 tcp]$ mpirun -np 2 -host bend001 -report-bindings -mca 
rmaps_lama_bind 1c -mca rmaps lama hostname
bend001
[bend001:17005] MCW rank 1 bound to socket 0[core 1[hwt 0-1]]: 
[../BB/../../../..][../../../../../..]
[bend001:17005] MCW rank 0 bound to socket 0[core 0[hwt 0-1]]: 
[BB/../../../../..][../../../../../..]
bend001
[rhc@bend001 tcp]$ 

(I also checked the internals using "-mca rmaps_base_verbose 10") so it could 
be your hier inversion causing problems again. Or it could be that you are 
hitting a connection issue we are seeing in some scenarios in the OOB subsystem 
- though if you are able to run using a non-lama mapper, that would seem 
unlikely.


On Dec 20, 2013, at 8:09 PM, tmish...@jcity.maeda.co.jp wrote:

> 
> 
> Hi Ralph,
> 
> Thank you very much. I tried many things such as:
> 
> mpirun -np 2 -host node05 -report-bindings -mca rmaps lama -mca
> rmaps_lama_bind 1c myprog
> 
> But every try failed. At least they were accepted by openmpi-1.7.3 as far
> as I remember.
> Anyway, please check it when you have a time, because using lama comes from
> my curiosity.
> 
> Regards,
> Tetsuya Mishima
> 
> 
>> I'll try to take a look at it - my expectation is that lama might get
> stuck because you didn't tell it a pattern to map, and I doubt that code
> path has seen much testing.
>> 
>> 
>> On Dec 20, 2013, at 5:52 PM, tmish...@jcity.maeda.co.jp wrote:
>> 
>>> 
>>> 
>>> Hi Ralph, I'm glad to hear that, thanks.
>>> 
>>> By the way, yesterday I tried to check how lama in 1.7.4rc treat numa
> node.
>>> 
>>> Then, even wiht this simple command line, it freezed without any
> massage:
>>> 
>>> mpirun -np 2 -host node05 -mca rmaps lama myprog
>>> 
>>> Could you check what happened?
>>> 
>>> Is it better to open new thread or continue this thread?
>>> 
>>> Regards,
>>> Tetsuya Mishima
>>> 
>>> 
>>>> I'll make it work so that NUMA can be either above or below socket
>>>> 
>>>> On Dec 20, 2013, at 2:57 AM, tmish...@jcity.maeda.co.jp wrote:
>>>> 
>>>>> 
>>>>> 
>>>>> Hi Brice,
>>>>> 
>>>>> Thank you for your comment. I understand what you mean.
>>>>> 
>>>>> My opinion was made just considering easy way to adjust the code for
>>>>> inversion of hierarchy in object tree.
>>>>> 
>>>>> Tetsuya Mishima
>>>>> 
>>>>> 
>>>>>> I don't think there's any such difference.
>>>>>> Also, all these NUMA architectures are reported the same by hwloc,
> and
>>>>>> therefore used the same in Open MPI.
>>>>>> 
>>>>>> And yes, L3 and NUMA are topologically-identical on AMD Magny-Cours
>>> (and
>>>>>> most recent AMD and Intel platforms).
>>>>>> 
>>>>>> Brice
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> Le 20/12/2013 11:33, tmish...@jcity.maeda.co.jp a écrit :
>>>>>>> 
>>>>>>> Hi Ralph,
>>>>>>> 
>>>>>>> The numa-node in AMD Mangy-Cours/Interlagos is so called cc(cache
>>>>>>> coherent)NUMA,
>>>>>>> which seems to be a little bit different from the traditional numa
>>>>> defined
>>>>>>> in openmpi.
>>>>>>> 
>>>>>>> I notice that ccNUMA object is almost same as L3cache object.
>>>>>>> So "-bind-to l3cache" or "-map-by l3cache" is valid for what I want
>>> to
>>>>> do.
>>>>>>> Therefore, "do not touch it" is one of the solution, I think ...
>>>>>>> 
>>>>>>> Anyway, mixing up these two types of numa is the problem.
>>>>>>> 
>>>>>>> Regards,
>>>>>>> Tetsuya Mishima
>>>>>>> 
>>>>>>>> I can wait it'll be fixed in 1.7.5 or later, because putting
>>> "-bind-to
>>>>>>>> numa"
>>>>>>>> and "-map-by numa" at the same time works as a workaround.
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> Tetsuya Mishima
>>>>>>>> 
>>>>>>>>> Yeah, it will impact everything that uses hwloc topology maps, I
>>>>> fear.
>>>>>>>>> 
>>>>>>>>> One side note: you'll need to add --hetero-nodes to your cmd
> line.
>>> If
>>>>>>> we
>>>>>>>> don't see that, we assume that all the node topologies are
> identical
>>> -
>>>>>>>> which clearly isn't true here.
>>>>>>>>> I'll try to resolve the hier inversion over the holiday - won't
> be
>>>>> for
>>>>>>>> 1.7.4, but hopefully for 1.7.5
>>>>>>>>> Thanks
>>>>>>>>> Ralph
>>>>>>>>> 
>>>>>>>>> On Dec 18, 2013, at 9:44 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> I think it's normal for AMD opteron having 8/16 cores such as
>>>>>>>>>> magny cours or interlagos. Because it usually has 2 numa nodes
>>>>>>>>>> in a cpu(socket), numa-node can not include a socket. This type
>>>>>>>>>> of hierarchy would be natural.
>>>>>>>>>> 
>>>>>>>>>> (node03 is Dell PowerEdge R815 and maybe quite common, I guess)
>>>>>>>>>> 
>>>>>>>>>> By the way, I think this inversion should affect rmaps_lama
>>> mapping.
>>>>>>>>>> 
>>>>>>>>>> Tetsuya Mishima
>>>>>>>>>> 
>>>>>>>>>>> Ick - yeah, that would be a problem. I haven't seen that type
> of
>>>>>>>>>> hierarchical inversion before - is node03 a different type of
>>> chip?
>>>>>>>>>>> Might take awhile for me to adjust the code to handle hier
>>>>>>>>>> inversion... :-(
>>>>>>>>>>> On Dec 18, 2013, at 9:05 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> Hi Ralph,
>>>>>>>>>>>> 
>>>>>>>>>>>> I found the reason. I attached the main part of output with 32
>>>>>>>>>>>> core node(node03) and 8 core node(node05) at the bottom.
>>>>>>>>>>>> 
>>>>>>>>>>>> From this information, socket of node03 includes numa-node.
>>>>>>>>>>>> On the other hand, numa-node of node05 includes socket.
>>>>>>>>>>>> The direction of object tree is opposite.
>>>>>>>>>>>> 
>>>>>>>>>>>> Since "-map-by socket" may be assumed as default,
>>>>>>>>>>>> for node05, "-bind-to numa and -map-by socket" means
>>>>>>>>>>>> upward search. For node03, this should be downward.
>>>>>>>>>>>> 
>>>>>>>>>>>> I guess that openmpi-1.7.4rc1 will always assume numa-node
>>>>>>>>>>>> includes socket. Is it right? Then, upward search is assumed
>>>>>>>>>>>> in orte_rmaps_base_compute_bindings even for node03 when I
>>>>>>>>>>>> put "-bind-to numa and -map-by socket" option.
>>>>>>>>>>>> 
>>>>>>>>>>>> [node03.cluster:15508] [[38286,0],0] rmaps:base:compute_usage
>>>>>>>>>>>> [node03.cluster:15508] mca:rmaps: compute bindings for job
>>>>>>> [38286,1]
>>>>>>>>>> with
>>>>>>>>>>>> policy NUMA
>>>>>>>>>>>> [node03.cluster:15508] mca:rmaps: bind upwards for job
> [38286,1]
>>>>>>> with
>>>>>>>>>>>> bindings NUMA
>>>>>>>>>>>> [node03.cluster:15508] [[38286,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> Machine
>>>>>>>>>>>> 
>>>>>>>>>>>> That's the reason of this trouble. Therefore, adding "-map-by
>>>>> core"
>>>>>>>>>> works.
>>>>>>>>>>>> (mapping pattern seems to be strange ...)
>>>>>>>>>>>> 
>>>>>>>>>>>> [mishima@node03 demos]$ mpirun -np 8 -bind-to numa -map-by
> core
>>>>>>>>>>>> -report-bindings myprog
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>> Cache
>>>>>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target
> NUMANode
>>>>>>> type
>>>>>>>>>>>> NUMANode
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 2 bound to socket 0[core 0[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 3 bound to socket 0[core 0[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 4 bound to socket 0[core 4[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 5 bound to socket 0[core 4[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 6 bound to socket 0[core 4[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 7 bound to socket 0[core 4[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 0 bound to socket 0[core 0[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> [node03.cluster:15885] MCW rank 1 bound to socket 0[core 0[hwt
>>>>> 0]],
>>>>>>>>>> socket
>>>>>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>>>>>> 
>>>>>>> 
> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>> Hello world from process 6 of 8
>>>>>>>>>>>> Hello world from process 5 of 8
>>>>>>>>>>>> Hello world from process 0 of 8
>>>>>>>>>>>> Hello world from process 7 of 8
>>>>>>>>>>>> Hello world from process 3 of 8
>>>>>>>>>>>> Hello world from process 4 of 8
>>>>>>>>>>>> Hello world from process 2 of 8
>>>>>>>>>>>> Hello world from process 1 of 8
>>>>>>>>>>>> 
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> Tetsuya Mishima
>>>>>>>>>>>> 
>>>>>>>>>>>> [node03.cluster:15508] Type: Machine Number of child objects:
> 4
>>>>>>>>>>>>    Name=NULL
>>>>>>>>>>>>    total=132358820KB
>>>>>>>>>>>>    Backend=Linux
>>>>>>>>>>>>    OSName=Linux
>>>>>>>>>>>>    OSRelease=2.6.18-308.16.1.el5
>>>>>>>>>>>>    OSVersion="#1 SMP Tue Oct 2 22:01:43 EDT 2012"
>>>>>>>>>>>>    Architecture=x86_64
>>>>>>>>>>>>    Cpuset:  0xffffffff
>>>>>>>>>>>>    Online:  0xffffffff
>>>>>>>>>>>>    Allowed: 0xffffffff
>>>>>>>>>>>>    Bind CPU proc:   TRUE
>>>>>>>>>>>>    Bind CPU thread: TRUE
>>>>>>>>>>>>    Bind MEM proc:   FALSE
>>>>>>>>>>>>    Bind MEM thread: TRUE
>>>>>>>>>>>>    Type: Socket Number of child objects: 2
>>>>>>>>>>>>            Name=NULL
>>>>>>>>>>>>            total=33071780KB
>>>>>>>>>>>>            CPUModel="AMD Opteron(tm) Processor 6136"
>>>>>>>>>>>>            Cpuset:  0x000000ff
>>>>>>>>>>>>            Online:  0x000000ff
>>>>>>>>>>>>            Allowed: 0x000000ff
>>>>>>>>>>>>            Type: NUMANode Number of child objects: 1
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> [node05.cluster:21750] Type: Machine Number of child objects:
> 2
>>>>>>>>>>>>    Name=NULL
>>>>>>>>>>>>    total=33080072KB
>>>>>>>>>>>>    Backend=Linux
>>>>>>>>>>>>    OSName=Linux
>>>>>>>>>>>>    OSRelease=2.6.18-308.16.1.el5
>>>>>>>>>>>>    OSVersion="#1 SMP Tue Oct 2 22:01:43 EDT 2012"
>>>>>>>>>>>>    Architecture=x86_64
>>>>>>>>>>>>    Cpuset:  0x000000ff
>>>>>>>>>>>>    Online:  0x000000ff
>>>>>>>>>>>>    Allowed: 0x000000ff
>>>>>>>>>>>>    Bind CPU proc:   TRUE
>>>>>>>>>>>>    Bind CPU thread: TRUE
>>>>>>>>>>>>    Bind MEM proc:   FALSE
>>>>>>>>>>>>    Bind MEM thread: TRUE
>>>>>>>>>>>>    Type: NUMANode Number of child objects: 1
>>>>>>>>>>>>            Name=NULL
>>>>>>>>>>>>            local=16532232KB
>>>>>>>>>>>>            total=16532232KB
>>>>>>>>>>>>            Cpuset:  0x0000000f
>>>>>>>>>>>>            Online:  0x0000000f
>>>>>>>>>>>>            Allowed: 0x0000000f
>>>>>>>>>>>>            Type: Socket Number of child objects: 1
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>>> Hmm...try adding "-mca rmaps_base_verbose 10 -mca
>>>>> ess_base_verbose
>>>>>>>> 5"
>>>>>>>>>> to
>>>>>>>>>>>> your cmd line and let's see what it thinks it found.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On Dec 18, 2013, at 6:55 PM, tmish...@jcity.maeda.co.jp
> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi, I report one more problem with openmpi-1.7.4rc1,
>>>>>>>>>>>>>> which is more serious.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> For our 32 core nodes(AMD magny cours based) which has
>>>>>>>>>>>>>> 8 numa-nodes, "-bind-to numa" does not work. Without
>>>>>>>>>>>>>> this option, it works. For your infomation, at the
>>>>>>>>>>>>>> bottom of this mail, I added the lstopo information
>>>>>>>>>>>>>> of the node.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>> Tetsuya Mishima
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> [mishima@manage ~]$ qsub -I -l nodes=1:ppn=32
>>>>>>>>>>>>>> qsub: waiting for job 8352.manage.cluster to start
>>>>>>>>>>>>>> qsub: job 8352.manage.cluster ready
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> [mishima@node03 demos]$ mpirun -np 8 -report-bindings
> -bind-to
>>>>>>> numa
>>>>>>>>>>>> myprog
>>>>>>>>>>>>>> [node03.cluster:15316] [[37582,0],0] bind:upward target
>>> NUMANode
>>>>>>>> type
>>>>>>>>>>>>>> Machine
>>>>>>>>>>>>>> 
>>>>>>> 
>>>>> 
>>> 
> --------------------------------------------------------------------------
>>>>>>>>>>>>>> A request was made to bind to NUMA, but an appropriate
> target
>>>>>>> could
>>>>>>>>>> not
>>>>>>>>>>>>>> be found on node node03.
>>>>>>>>>>>>>> 
>>>>>>> 
>>>>> 
>>> 
> --------------------------------------------------------------------------
>>>>>>>>>>>>>> [mishima@node03 ~]$ cd ~/Desktop/openmpi-1.7/demos/
>>>>>>>>>>>>>> [mishima@node03 demos]$ mpirun -np 8 -report-bindings myprog
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 2 bound to socket 1[core 8
> [hwt
>>>>>>> 0]]:
>>>>>>>>>>>>>> [./././././././.][B/././././././.][./././././././.][
>>>>>>>>>>>>>> ./././././././.]>>>>>>>>>>>> [node03.cluster:15282] MCW rank
> 3 bound to socket 1[core 9[hwt
>>>>>>> 0]]:
>>>>>>>>>>>>>> [./././././././.][./B/./././././.][./././././././.][
>>>>>>>>>>>>>> ./././././././.]
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 4 bound to socket 2[core 16
>>> [hwt
>>>>>>>> 0]]:
>>>>>>>>>>>>>> [./././././././.][./././././././.][B/././././././.]
>>>>>>>>>>>>>> [./././././././.]
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 5 bound to socket 2[core 17
>>> [hwt
>>>>>>>> 0]]:
>>>>>>>>>>>>>> [./././././././.][./././././././.][./B/./././././.]
>>>>>>>>>>>>>> [./././././././.]
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 6 bound to socket 3[core 24
>>> [hwt
>>>>>>>> 0]]:
>>>>>>>>>>>>>> [./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>>>> [B/././././././.]
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 7 bound to socket 3[core 25
>>> [hwt
>>>>>>>> 0]]:
>>>>>>>>>>>>>> [./././././././.][./././././././.][./././././././.]
>>>>>>>>>>>>>> [./B/./././././.]
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 0 bound to socket 0[core 0
> [hwt
>>>>>>> 0]]:
>>>>>>>>>>>>>> [B/././././././.][./././././././.][./././././././.][
>>>>>>>>>>>>>> ./././././././.]
>>>>>>>>>>>>>> [node03.cluster:15282] MCW rank 1 bound to socket 0[core 1
> [hwt
>>>>>>> 0]]:
>>>>>>>>>>>>>> [./B/./././././.][./././././././.][./././././././.][
>>>>>>>>>>>>>> ./././././././.]
>>>>>>>>>>>>>> Hello world from process 2 of 8
>>>>>>>>>>>>>> Hello world from process 5 of 8
>>>>>>>>>>>>>> Hello world from process 4 of 8
>>>>>>>>>>>>>> Hello world from process 3 of 8>>>>>>>>>> Hello world from
>>> process 1 of 8
>>>>>>>>>>>>>> Hello world from process 7 of 8
>>>>>>>>>>>>>> Hello world from process 6 of 8
>>>>>>>>>>>>>> Hello world from process 0 of 8
>>>>>>>>>>>>>> [mishima@node03 demos]$ ~/opt/hwloc/bin/lstopo-no-graphics
>>>>>>>>>>>>>> Machine (126GB)
>>>>>>>>>>>>>> Socket L#0 (32GB)
>>>>>>>>>>>>>> NUMANode L#0 (P#0 16GB) + L3 L#0 (5118KB)
>>>>>>>>>>>>>> L2 L#0 (512KB) + L1d L#0 (64KB) + L1i L#0 (64KB) + Core L#0
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#0
>>>>>>>>>>>>>> (P#0)
>>>>>>>>>>>>>> L2 L#1 (512KB) + L1d L#1 (64KB) + L1i L#1 (64KB) + Core L#1
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#1
>>>>>>>>>>>>>> (P#1)
>>>>>>>>>>>>>> L2 L#2 (512KB) + L1d L#2 (64KB) + L1i L#2 (64KB) + Core L#2
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#2
>>>>>>>>>>>>>> (P#2)
>>>>>>>>>>>>>> L2 L#3 (512KB) + L1d L#3 (64KB) + L1i L#3 (64KB) + Core L#3
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#3
>>>>>>>>>>>>>> (P#3)
>>>>>>>>>>>>>> NUMANode L#1 (P#1 16GB) + L3 L#1 (5118KB)
>>>>>>>>>>>>>> L2 L#4 (512KB) + L1d L#4 (64KB) + L1i L#4 (64KB) + Core L#4
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#4
>>>>>>>>>>>>>> (P#4)
>>>>>>>>>>>>>> L2 L#5 (512KB) + L1d L#5 (64KB) + L1i L#5 (64KB) + Core L#5
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#5
>>>>>>>>>>>>>> (P#5)
>>>>>>>>>>>>>> L2 L#6 (512KB) + L1d L#6 (64KB) + L1i L#6 (64KB) + Core L#6
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#6
>>>>>>>>>>>>>> (P#6)
>>>>>>>>>>>>>> L2 L#7 (512KB) + L1d L#7 (64KB) + L1i L#7 (64KB) + Core L#7
>>> +
>>>>>>> PU>>>>>> L#7
>>>>>>>>>>>>>> (P#7)
>>>>>>>>>>>>>> Socket L#1 (32GB)
>>>>>>>>>>>>>> NUMANode L#2 (P#6 16GB) + L3 L#2 (5118KB)
>>>>>>>>>>>>>> L2 L#8 (512KB) + L1d L#8 (64KB) + L1i L#8 (64KB) + Core L#8
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#8
>>>>>>>>>>>>>> (P#8)
>>>>>>>>>>>>>> L2 L#9 (512KB) + L1d L#9 (64KB) + L1i L#9 (64KB) + Core L#9
>>> +
>>>>>>> PU
>>>>>>>>>>>> L#9
>>>>>>>>>>>>>> (P#9)
>>>>>>>>>>>>>> L2 L#10 (512KB) + L1d L#10 (64KB) + L1i L#10 (64KB) + Core
>>>>>>> L#10
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#10 (P#10)
>>>>>>>>>>>>>> L2 L#11 (512KB) + L1d L#11 (64KB) + L1i L#11 (64KB) + Core
>>>>>>> L#11
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#11 (P#11)
>>>>>>>>>>>>>> NUMANode L#3 (P#7 16GB) + L3 L#3 (5118KB)
>>>>>>>>>>>>>> L2 L#12 (512KB) + L1d L#12 (64KB) + L1i L#12 (64KB) + Core
>>>>>>> L#12
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#12 (P#12)
>>>>>>>>>>>>>> L2 L#13 (512KB) + L1d L#13 (64KB) + L1i L#13 (64KB) + Core
>>>>>>> L#13
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#13 (P#13)
>>>>>>>>>>>>>> L2 L#14 (512KB) + L1d L#14 (64KB) + L1i L#14 (64KB) + Core
>>>>>>> L#14
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#14 (P#14)
>>>>>>>>>>>>>> L2 L#15 (512KB) + L1d L#15 (64KB) + L1i L#15 (64KB) + Core
>>>>>>> L#15
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#15 (P#15)
>>>>>>>>>>>>>> Socket L#2 (32GB)
>>>>>>>>>>>>>> NUMANode L#4 (P#4 16GB) + L3 L#4 (5118KB)
>>>>>>>>>>>>>> L2 L#16 (512KB) + L1d L#16 (64KB) + L1i L#16 (64KB) + Core
>>>>>>> L#16
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#16 (P#16)
>>>>>>>>>>>>>> L2 L#17 (512KB) + L1d L#17 (64KB) + L1i L#17 (64KB) + Core
>>>>>>> L#17
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#17 (P#17)> >>>>>    L2 L#18 (512KB) + L1d L#18 (64KB) +
> L1i
>>>>>>> L#18 (64KB) + Core L#18
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#18 (P#18)
>>>>>>>>>>>>>> L2 L#19 (512KB) + L1d L#19 (64KB) + L1i L#19 (64KB) + Core
>>>>>>> L#19
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#19 (P#19)
>>>>>>>>>>>>>> NUMANode L#5 (P#5 16GB) + L3 L#5 (5118KB)
>>>>>>>>>>>>>> L2 L#20 (512KB) + L1d L#20 (64KB) + L1i L#20 (64KB) + Core
>>>>>>> L#20
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#20 (P#20)
>>>>>>>>>>>>>> L2 L#21 (512KB) + L1d L#21 (64KB) + L1i L#21 (64KB) + Core
>>>>>>> L#21
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#21 (P#21)
>>>>>>>>>>>>>> L2 L#22 (512KB) + L1d L#22 (64KB) + L1i L#22 (64KB) + Core
>>>>>>> L#22
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#22 (P#22)
>>>>>>>>>>>>>> L2 L#23 (512KB) + L1d L#23 (64KB) + L1i L#23 (64KB) + Core
>>>>>>> L#23
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#23 (P#23)
>>>>>>>>>>>>>> Socket L#3 (32GB)
>>>>>>>>>>>>>> NUMANode L#6 (P#2 16GB) + L3 L#6 (5118KB)
>>>>>>>>>>>>>> L2 L#24 (512KB) + L1d L#24 (64KB) + L1i L#24 (64KB) + Core
>>>>>>> L#24
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#24 (P#24)>>>>>    L2 L#25 (512KB) + L1d L#25 (64KB) + L1i
>>> L#25
>>>>>>>> (64KB) + Core L#25 +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#25 (P#25)
>>>>>>>>>>>>>> L2 L#26 (512KB) + L1d L#26 (64KB) + L1i L#26 (64KB) + Core
>>>>>>> L#26
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#26 (P#26)
>>>>>>>>>>>>>> L2 L#27 (512KB) + L1d L#27 (64KB) + L1i L#27 (64KB) + Core
>>>>>>> L#27
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#27 (P#27)
>>>>>>>>>>>>>> NUMANode L#7 (P#3 16GB) + L3 L#7 (5118KB)
>>>>>>>>>>>>>> L2 L#28 (512KB) + L1d L#28 (64KB) + L1i L#28 (64KB) + Core
>>>>>>> L#28
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#28 (P#28)
>>>>>>>>>>>>>> L2 L#29 (512KB) + L1d L#29 (64KB) + L1i L#29 (64KB) + Core
>>>>>>> L#29
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#29 (P#29)
>>>>>>>>>>>>>> L2 L#30 (512KB) + L1d L#30 (64KB) + L1i L#30 (64KB) + Core
>>>>>>> L#30
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#30 (P#30)
>>>>>>>>>>>>>> L2 L#31 (512KB) + L1d L#31 (64KB) + L1i L#31 (64KB) + Core
>>>>>>> L#31
>>>>>>>> +
>>>>>>>>>>>> PU
>>>>>>>>>>>>>> L#31 (P#31)
>>>>>>>>>>>>>> HostBridge L#0
>>>>>>>>>>>>>> PCIBridge
>>>>>>>>>>>>>> PCI 14e4:1639
>>>>>>>>>>>>>>   Net L#0 "eth0"
>>>>>>>>>>>>>> PCI 14e4:1639
>>>>>>>>>>>>>>   Net L#1 "eth1"
>>>>>>>>>>>>>> PCIBridge
>>>>>>>>>>>>>> PCI 14e4:1639
>>>>>>>>>>>>>>   Net L#2 "eth2"
>>>>>>>>>>>>>> PCI 14e4:1639
>>>>>>>>>>>>>>   Net L#3 "eth3"
>>>>>>>>>>>>>> PCIBridge
>>>>>>>>>>>>>> PCIBridge
>>>>>>>>>>>>>>   PCIBridge
>>>>>>>>>>>>>>     PCI 1000:0072
>>>>>>>>>>>>>>       Block L#4 "sdb"
>>>>>>>>>>>>>>       Block L#5 "sda"
>>>>>>>>>>>>>> PCI 1002:4390
>>>>>>>>>>>>>> Block L#6 "sr0"
>>>>>>>>>>>>>> PCIBridge
>>>>>>>>>>>>>> PCI 102b:0532
>>>>>>>>>>>>>> HostBridge L#7
>>>>>>>>>>>>>> PCIBridge
>>>>>>>>>>>>>> PCI 15b3:6274
>>>>>>>>>>>>>>   Net L#7 "ib0"
>>>>>>>>>>>>>>   OpenFabrics L#8 "mthca0"
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>> us...@open-mpi.org>>
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> users mailing list
>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to