Re: [OMPI users] usNIC point-to-point messaging module

2014-04-04 Thread Ralph Castain
Fixed in r31308 and scheduled for inclusion in 1.8.1

Thanks
Ralph

On Apr 2, 2014, at 12:17 PM, Ralph Castain  wrote:

> Yeah, it's a change we added to resolve a problem when Slurm is configured 
> with TaskAffinity set. It's harmless, but annoying - I'm trying to figure out 
> a solution.
> 
> 
> 
> On Wed, Apr 2, 2014 at 11:35 AM, Dave Goodell (dgoodell)  
> wrote:
> On Apr 2, 2014, at 12:57 PM, Filippo Spiga  wrote:
> 
> > I still do not understand why this keeps appearing...
> >
> > srun: cluster configuration lacks support for cpu binding
> >
> > Any clue?
> 
> I don't know what causes that message.  Ralph, any thoughts here?
> 
> -Dave
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 



Re: [OMPI users] usNIC point-to-point messaging module

2014-04-02 Thread Ralph Castain
Yeah, it's a change we added to resolve a problem when Slurm is configured
with TaskAffinity set. It's harmless, but annoying - I'm trying to figure
out a solution.



On Wed, Apr 2, 2014 at 11:35 AM, Dave Goodell (dgoodell)  wrote:

> On Apr 2, 2014, at 12:57 PM, Filippo Spiga 
> wrote:
>
> > I still do not understand why this keeps appearing...
> >
> > srun: cluster configuration lacks support for cpu binding
> >
> > Any clue?
>
> I don't know what causes that message.  Ralph, any thoughts here?
>
> -Dave
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Re: [OMPI users] usNIC point-to-point messaging module

2014-04-02 Thread Dave Goodell (dgoodell)
On Apr 2, 2014, at 12:57 PM, Filippo Spiga  wrote:

> I still do not understand why this keeps appearing...
> 
> srun: cluster configuration lacks support for cpu binding
> 
> Any clue?

I don't know what causes that message.  Ralph, any thoughts here?

-Dave



Re: [OMPI users] usNIC point-to-point messaging module

2014-04-02 Thread Filippo Spiga
Dear Dave,

your suggestion worked, the file was there ar delete the directory and rebuild 
everything solved the issue. I still do not understand why this keeps 
appearing...

srun: cluster configuration lacks support for cpu binding

Any clue?

F


On Apr 1, 2014, at 9:08 PM, Dave Goodell (dgoodell)  wrote:

> On Apr 1, 2014, at 12:13 PM, Filippo Spiga  wrote:
> 
>> Dear Ralph, Dear Jeff,
>> 
>> I've just recompiled the latest Open MPI 1.8. I added 
>> "--enable-mca-no-build=btl-usnic" to configure but the message still appear. 
>> Here the output of "--mca btl_base_verbose 100" (trunked immediately after 
>> the application starts)
> 
> Jeff's on vacation, so I'll see if I can help here.
> 
> Try deleting all the files in "$PREFIX/lib/openmpi/", where "$PREFIX" is the 
> value you passed to configure with "--prefix=".  If you did not pass a value, 
> then it is "/usr/local".  Then reinstall (with "make install" in the OMPI 
> build tree).
> 
> What I think is happening is that you still have an "mca_btl_usnic.so" file 
> leftover from the last time you installed OMPI (before passing 
> "--enable-mca-no-build=btl-usnic").  So OMPI is using this shared library and 
> you get exactly the same problem.
> 
> -Dave
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

--
Mr. Filippo SPIGA, M.Sc.
http://www.linkedin.com/in/filippospiga ~ skype: filippo.spiga

«Nobody will drive us out of Cantor's paradise.» ~ David Hilbert

*
Disclaimer: "Please note this message and any attachments are CONFIDENTIAL and 
may be privileged or otherwise protected from disclosure. The contents are not 
to be disclosed to anyone other than the addressee. Unauthorized recipients are 
requested to preserve this confidentiality and to advise the sender immediately 
of any error in transmission."




Re: [OMPI users] usNIC point-to-point messaging module

2014-04-01 Thread Dave Goodell (dgoodell)
On Apr 1, 2014, at 12:13 PM, Filippo Spiga  wrote:

> Dear Ralph, Dear Jeff,
> 
> I've just recompiled the latest Open MPI 1.8. I added 
> "--enable-mca-no-build=btl-usnic" to configure but the message still appear. 
> Here the output of "--mca btl_base_verbose 100" (trunked immediately after 
> the application starts)

Jeff's on vacation, so I'll see if I can help here.

Try deleting all the files in "$PREFIX/lib/openmpi/", where "$PREFIX" is the 
value you passed to configure with "--prefix=".  If you did not pass a value, 
then it is "/usr/local".  Then reinstall (with "make install" in the OMPI build 
tree).

What I think is happening is that you still have an "mca_btl_usnic.so" file 
leftover from the last time you installed OMPI (before passing 
"--enable-mca-no-build=btl-usnic").  So OMPI is using this shared library and 
you get exactly the same problem.

-Dave



Re: [OMPI users] usNIC point-to-point messaging module

2014-04-01 Thread Filippo Spiga
Dear Ralph, Dear Jeff,

I've just recompiled the latest Open MPI 1.8. I added 
"--enable-mca-no-build=btl-usnic" to configure but the message still appear. 
Here the output of "--mca btl_base_verbose 100" (trunked immediately after the 
application starts)


srun: cluster configuration lacks support for cpu binding
[tesla88:26769] mca: base: components_register: registering btl components
[tesla88:26769] mca: base: components_register: found loaded component openib
[tesla88:26768] mca: base: components_register: registering btl components
[tesla88:26768] mca: base: components_register: found loaded component openib
[tesla88:26768] mca: base: components_register: component openib register 
function successful
[tesla88:26769] mca: base: components_register: component openib register 
function successful
[tesla88:26769] mca: base: components_register: found loaded component self
[tesla88:26768] mca: base: components_register: found loaded component self
[tesla88:26769] mca: base: components_register: component self register 
function successful
[tesla88:26768] mca: base: components_register: component self register 
function successful
[tesla88:26769] mca: base: components_register: found loaded component sm
[tesla88:26768] mca: base: components_register: found loaded component sm
[tesla88:26769] mca: base: components_register: component sm register function 
successful
[tesla88:26768] mca: base: components_register: component sm register function 
successful
[tesla88:26769] mca: base: components_register: found loaded component tcp
[tesla88:26768] mca: base: components_register: found loaded component tcp
[tesla88:26769] mca: base: components_register: component tcp register function 
successful
[tesla88:26768] mca: base: components_register: component tcp register function 
successful
[tesla88:26769] mca: base: components_register: found loaded component usnic
[tesla88:26768] mca: base: components_register: found loaded component usnic
[tesla88:26769] mca: base: components_register: component usnic register 
function successful
[tesla88:26768] mca: base: components_register: component usnic register 
function successful
[tesla88:26769] mca: base: components_register: found loaded component vader
[tesla88:26768] mca: base: components_register: found loaded component vader
[tesla88:26769] mca: base: components_register: component vader register 
function successful
[tesla88:26769] mca: base: components_open: opening btl components
[tesla88:26769] mca: base: components_open: found loaded component openib
[tesla88:26769] mca: base: components_open: component openib open function 
successful
[tesla88:26769] mca: base: components_open: found loaded component self
[tesla88:26768] mca: base: components_register: component vader register 
function successful
[tesla88:26769] mca: base: components_open: component self open function 
successful
[tesla88:26769] mca: base: components_open: found loaded component sm
[tesla88:26769] mca: base: components_open: component sm open function 
successful
[tesla88:26769] mca: base: components_open: found loaded component tcp
[tesla88:26768] mca: base: components_open: opening btl components
[tesla88:26768] mca: base: components_open: found loaded component openib
[tesla88:26768] mca: base: components_open: component openib open function 
successful
[tesla88:26768] mca: base: components_open: found loaded component self
[tesla88:26768] mca: base: components_open: component self open function 
successful
[tesla88:26768] mca: base: components_open: found loaded component sm
[tesla88:26768] mca: base: components_open: component sm open function 
successful
[tesla88:26768] mca: base: components_open: found loaded component tcp
[tesla88:26769] mca: base: components_open: component tcp open function 
successful
[tesla88:26769] mca: base: components_open: found loaded component usnic
[tesla88:26769] mca: base: components_open: component usnic open function 
successful
[tesla88:26769] mca: base: components_open: found loaded component vader
[tesla88:26769] mca: base: components_open: component vader open function 
successful
[tesla89:45456] mca: base: components_register: registering btl components
[tesla89:45456] mca: base: components_register: found loaded component openib
[tesla88:26768] mca: base: components_open: component tcp open function 
successful
[tesla88:26768] mca: base: components_open: found loaded component usnic
[tesla88:26768] mca: base: components_open: component usnic open function 
successful
[tesla88:26768] mca: base: components_open: found loaded component vader
[tesla88:26768] mca: base: components_open: component vader open function 
successful
[tesla89:45455] mca: base: components_register: registering btl components
[tesla89:45455] mca: base: components_register: found loaded component openib
[tesla89:45456] mca: base: components_register: component openib register 
function successful
[tesla89:45456] mca: base: components_register: found loaded component self

Re: [OMPI users] usNIC point-to-point messaging module

2014-03-24 Thread Jeff Squyres (jsquyres)
No, this is not a configure issue -- the usnic BTL uses the verbs API.

The usnic BTL should be disqualifying itself at runtime, though, if you don't 
have usNIC devices.

Are you running on Cisco UCS servers with Cisco VICs, perchance?

If not, could you send the output of "mpirun --mca btl_base_verbose 100 ..."?



On Mar 23, 2014, at 10:35 AM, Ralph Castain  wrote:

> Hmmm...we'll have to check the configure logic as I don't think you should be 
> getting that message. Regardless, it isn't something of concern - you can 
> turn it "off" by adding
> 
> -mca btl ^usnic
> 
> on your command line, or configuring OMPI --enable-mca-no-build=btl-usnic
> 
> 
> On Mar 22, 2014, at 10:00 PM, Filippo Spiga  wrote:
> 
>> Dear all,
>> 
>> I recompiled Open MPI 1.7.5 a couple of time, I am sure I have been selected 
>> openib. However I have some doubts because this message
>> 
>> --
>> [[28098,1],8]: A high-performance Open MPI point-to-point messaging module
>> was unable to find any relevant network interfaces:
>> 
>> Module: usNIC
>> Host: tesla79
>> 
>> Another transport will be used instead, although this may result in
>> lower performance.
>> --
>> 
>> keeps popping up. I am really worried there might be a degradation of 
>> performance because of this warning. Any clue about where it came from and 
>> how I can let it disappear?
>> 
>> Thanks in advance,
>> Filippo
>> 
>> --
>> Mr. Filippo SPIGA, M.Sc.
>> http://www.linkedin.com/in/filippospiga ~ skype: filippo.spiga
>> 
>> «Nobody will drive us out of Cantor's paradise.» ~ David Hilbert
>> 
>> *
>> Disclaimer: "Please note this message and any attachments are CONFIDENTIAL 
>> and may be privileged or otherwise protected from disclosure. The contents 
>> are not to be disclosed to anyone other than the addressee. Unauthorized 
>> recipients are requested to preserve this confidentiality and to advise the 
>> sender immediately of any error in transmission."
>> 
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/



Re: [OMPI users] usNIC point-to-point messaging module

2014-03-23 Thread Ralph Castain
Hmmm...we'll have to check the configure logic as I don't think you should be 
getting that message. Regardless, it isn't something of concern - you can turn 
it "off" by adding

-mca btl ^usnic

on your command line, or configuring OMPI --enable-mca-no-build=btl-usnic


On Mar 22, 2014, at 10:00 PM, Filippo Spiga  wrote:

> Dear all,
> 
> I recompiled Open MPI 1.7.5 a couple of time, I am sure I have been selected 
> openib. However I have some doubts because this message
> 
> --
> [[28098,1],8]: A high-performance Open MPI point-to-point messaging module
> was unable to find any relevant network interfaces:
> 
> Module: usNIC
>  Host: tesla79
> 
> Another transport will be used instead, although this may result in
> lower performance.
> --
> 
> keeps popping up. I am really worried there might be a degradation of 
> performance because of this warning. Any clue about where it came from and 
> how I can let it disappear?
> 
> Thanks in advance,
> Filippo
> 
> --
> Mr. Filippo SPIGA, M.Sc.
> http://www.linkedin.com/in/filippospiga ~ skype: filippo.spiga
> 
> «Nobody will drive us out of Cantor's paradise.» ~ David Hilbert
> 
> *
> Disclaimer: "Please note this message and any attachments are CONFIDENTIAL 
> and may be privileged or otherwise protected from disclosure. The contents 
> are not to be disclosed to anyone other than the addressee. Unauthorized 
> recipients are requested to preserve this confidentiality and to advise the 
> sender immediately of any error in transmission."
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users