Re: [OMPI users] usNIC point-to-point messaging module
Fixed in r31308 and scheduled for inclusion in 1.8.1 Thanks Ralph On Apr 2, 2014, at 12:17 PM, Ralph Castainwrote: > Yeah, it's a change we added to resolve a problem when Slurm is configured > with TaskAffinity set. It's harmless, but annoying - I'm trying to figure out > a solution. > > > > On Wed, Apr 2, 2014 at 11:35 AM, Dave Goodell (dgoodell) > wrote: > On Apr 2, 2014, at 12:57 PM, Filippo Spiga wrote: > > > I still do not understand why this keeps appearing... > > > > srun: cluster configuration lacks support for cpu binding > > > > Any clue? > > I don't know what causes that message. Ralph, any thoughts here? > > -Dave > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >
Re: [OMPI users] usNIC point-to-point messaging module
Yeah, it's a change we added to resolve a problem when Slurm is configured with TaskAffinity set. It's harmless, but annoying - I'm trying to figure out a solution. On Wed, Apr 2, 2014 at 11:35 AM, Dave Goodell (dgoodell)wrote: > On Apr 2, 2014, at 12:57 PM, Filippo Spiga > wrote: > > > I still do not understand why this keeps appearing... > > > > srun: cluster configuration lacks support for cpu binding > > > > Any clue? > > I don't know what causes that message. Ralph, any thoughts here? > > -Dave > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >
Re: [OMPI users] usNIC point-to-point messaging module
On Apr 2, 2014, at 12:57 PM, Filippo Spigawrote: > I still do not understand why this keeps appearing... > > srun: cluster configuration lacks support for cpu binding > > Any clue? I don't know what causes that message. Ralph, any thoughts here? -Dave
Re: [OMPI users] usNIC point-to-point messaging module
Dear Dave, your suggestion worked, the file was there ar delete the directory and rebuild everything solved the issue. I still do not understand why this keeps appearing... srun: cluster configuration lacks support for cpu binding Any clue? F On Apr 1, 2014, at 9:08 PM, Dave Goodell (dgoodell)wrote: > On Apr 1, 2014, at 12:13 PM, Filippo Spiga wrote: > >> Dear Ralph, Dear Jeff, >> >> I've just recompiled the latest Open MPI 1.8. I added >> "--enable-mca-no-build=btl-usnic" to configure but the message still appear. >> Here the output of "--mca btl_base_verbose 100" (trunked immediately after >> the application starts) > > Jeff's on vacation, so I'll see if I can help here. > > Try deleting all the files in "$PREFIX/lib/openmpi/", where "$PREFIX" is the > value you passed to configure with "--prefix=". If you did not pass a value, > then it is "/usr/local". Then reinstall (with "make install" in the OMPI > build tree). > > What I think is happening is that you still have an "mca_btl_usnic.so" file > leftover from the last time you installed OMPI (before passing > "--enable-mca-no-build=btl-usnic"). So OMPI is using this shared library and > you get exactly the same problem. > > -Dave > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Mr. Filippo SPIGA, M.Sc. http://www.linkedin.com/in/filippospiga ~ skype: filippo.spiga «Nobody will drive us out of Cantor's paradise.» ~ David Hilbert * Disclaimer: "Please note this message and any attachments are CONFIDENTIAL and may be privileged or otherwise protected from disclosure. The contents are not to be disclosed to anyone other than the addressee. Unauthorized recipients are requested to preserve this confidentiality and to advise the sender immediately of any error in transmission."
Re: [OMPI users] usNIC point-to-point messaging module
On Apr 1, 2014, at 12:13 PM, Filippo Spigawrote: > Dear Ralph, Dear Jeff, > > I've just recompiled the latest Open MPI 1.8. I added > "--enable-mca-no-build=btl-usnic" to configure but the message still appear. > Here the output of "--mca btl_base_verbose 100" (trunked immediately after > the application starts) Jeff's on vacation, so I'll see if I can help here. Try deleting all the files in "$PREFIX/lib/openmpi/", where "$PREFIX" is the value you passed to configure with "--prefix=". If you did not pass a value, then it is "/usr/local". Then reinstall (with "make install" in the OMPI build tree). What I think is happening is that you still have an "mca_btl_usnic.so" file leftover from the last time you installed OMPI (before passing "--enable-mca-no-build=btl-usnic"). So OMPI is using this shared library and you get exactly the same problem. -Dave
Re: [OMPI users] usNIC point-to-point messaging module
Dear Ralph, Dear Jeff, I've just recompiled the latest Open MPI 1.8. I added "--enable-mca-no-build=btl-usnic" to configure but the message still appear. Here the output of "--mca btl_base_verbose 100" (trunked immediately after the application starts) srun: cluster configuration lacks support for cpu binding [tesla88:26769] mca: base: components_register: registering btl components [tesla88:26769] mca: base: components_register: found loaded component openib [tesla88:26768] mca: base: components_register: registering btl components [tesla88:26768] mca: base: components_register: found loaded component openib [tesla88:26768] mca: base: components_register: component openib register function successful [tesla88:26769] mca: base: components_register: component openib register function successful [tesla88:26769] mca: base: components_register: found loaded component self [tesla88:26768] mca: base: components_register: found loaded component self [tesla88:26769] mca: base: components_register: component self register function successful [tesla88:26768] mca: base: components_register: component self register function successful [tesla88:26769] mca: base: components_register: found loaded component sm [tesla88:26768] mca: base: components_register: found loaded component sm [tesla88:26769] mca: base: components_register: component sm register function successful [tesla88:26768] mca: base: components_register: component sm register function successful [tesla88:26769] mca: base: components_register: found loaded component tcp [tesla88:26768] mca: base: components_register: found loaded component tcp [tesla88:26769] mca: base: components_register: component tcp register function successful [tesla88:26768] mca: base: components_register: component tcp register function successful [tesla88:26769] mca: base: components_register: found loaded component usnic [tesla88:26768] mca: base: components_register: found loaded component usnic [tesla88:26769] mca: base: components_register: component usnic register function successful [tesla88:26768] mca: base: components_register: component usnic register function successful [tesla88:26769] mca: base: components_register: found loaded component vader [tesla88:26768] mca: base: components_register: found loaded component vader [tesla88:26769] mca: base: components_register: component vader register function successful [tesla88:26769] mca: base: components_open: opening btl components [tesla88:26769] mca: base: components_open: found loaded component openib [tesla88:26769] mca: base: components_open: component openib open function successful [tesla88:26769] mca: base: components_open: found loaded component self [tesla88:26768] mca: base: components_register: component vader register function successful [tesla88:26769] mca: base: components_open: component self open function successful [tesla88:26769] mca: base: components_open: found loaded component sm [tesla88:26769] mca: base: components_open: component sm open function successful [tesla88:26769] mca: base: components_open: found loaded component tcp [tesla88:26768] mca: base: components_open: opening btl components [tesla88:26768] mca: base: components_open: found loaded component openib [tesla88:26768] mca: base: components_open: component openib open function successful [tesla88:26768] mca: base: components_open: found loaded component self [tesla88:26768] mca: base: components_open: component self open function successful [tesla88:26768] mca: base: components_open: found loaded component sm [tesla88:26768] mca: base: components_open: component sm open function successful [tesla88:26768] mca: base: components_open: found loaded component tcp [tesla88:26769] mca: base: components_open: component tcp open function successful [tesla88:26769] mca: base: components_open: found loaded component usnic [tesla88:26769] mca: base: components_open: component usnic open function successful [tesla88:26769] mca: base: components_open: found loaded component vader [tesla88:26769] mca: base: components_open: component vader open function successful [tesla89:45456] mca: base: components_register: registering btl components [tesla89:45456] mca: base: components_register: found loaded component openib [tesla88:26768] mca: base: components_open: component tcp open function successful [tesla88:26768] mca: base: components_open: found loaded component usnic [tesla88:26768] mca: base: components_open: component usnic open function successful [tesla88:26768] mca: base: components_open: found loaded component vader [tesla88:26768] mca: base: components_open: component vader open function successful [tesla89:45455] mca: base: components_register: registering btl components [tesla89:45455] mca: base: components_register: found loaded component openib [tesla89:45456] mca: base: components_register: component openib register function successful [tesla89:45456] mca: base: components_register: found loaded component self
Re: [OMPI users] usNIC point-to-point messaging module
No, this is not a configure issue -- the usnic BTL uses the verbs API. The usnic BTL should be disqualifying itself at runtime, though, if you don't have usNIC devices. Are you running on Cisco UCS servers with Cisco VICs, perchance? If not, could you send the output of "mpirun --mca btl_base_verbose 100 ..."? On Mar 23, 2014, at 10:35 AM, Ralph Castainwrote: > Hmmm...we'll have to check the configure logic as I don't think you should be > getting that message. Regardless, it isn't something of concern - you can > turn it "off" by adding > > -mca btl ^usnic > > on your command line, or configuring OMPI --enable-mca-no-build=btl-usnic > > > On Mar 22, 2014, at 10:00 PM, Filippo Spiga wrote: > >> Dear all, >> >> I recompiled Open MPI 1.7.5 a couple of time, I am sure I have been selected >> openib. However I have some doubts because this message >> >> -- >> [[28098,1],8]: A high-performance Open MPI point-to-point messaging module >> was unable to find any relevant network interfaces: >> >> Module: usNIC >> Host: tesla79 >> >> Another transport will be used instead, although this may result in >> lower performance. >> -- >> >> keeps popping up. I am really worried there might be a degradation of >> performance because of this warning. Any clue about where it came from and >> how I can let it disappear? >> >> Thanks in advance, >> Filippo >> >> -- >> Mr. Filippo SPIGA, M.Sc. >> http://www.linkedin.com/in/filippospiga ~ skype: filippo.spiga >> >> «Nobody will drive us out of Cantor's paradise.» ~ David Hilbert >> >> * >> Disclaimer: "Please note this message and any attachments are CONFIDENTIAL >> and may be privileged or otherwise protected from disclosure. The contents >> are not to be disclosed to anyone other than the addressee. Unauthorized >> recipients are requested to preserve this confidentiality and to advise the >> sender immediately of any error in transmission." >> >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] usNIC point-to-point messaging module
Hmmm...we'll have to check the configure logic as I don't think you should be getting that message. Regardless, it isn't something of concern - you can turn it "off" by adding -mca btl ^usnic on your command line, or configuring OMPI --enable-mca-no-build=btl-usnic On Mar 22, 2014, at 10:00 PM, Filippo Spigawrote: > Dear all, > > I recompiled Open MPI 1.7.5 a couple of time, I am sure I have been selected > openib. However I have some doubts because this message > > -- > [[28098,1],8]: A high-performance Open MPI point-to-point messaging module > was unable to find any relevant network interfaces: > > Module: usNIC > Host: tesla79 > > Another transport will be used instead, although this may result in > lower performance. > -- > > keeps popping up. I am really worried there might be a degradation of > performance because of this warning. Any clue about where it came from and > how I can let it disappear? > > Thanks in advance, > Filippo > > -- > Mr. Filippo SPIGA, M.Sc. > http://www.linkedin.com/in/filippospiga ~ skype: filippo.spiga > > «Nobody will drive us out of Cantor's paradise.» ~ David Hilbert > > * > Disclaimer: "Please note this message and any attachments are CONFIDENTIAL > and may be privileged or otherwise protected from disclosure. The contents > are not to be disclosed to anyone other than the addressee. Unauthorized > recipients are requested to preserve this confidentiality and to advise the > sender immediately of any error in transmission." > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users