Good question. That isn't my decision. Michael, Should the ipmi plugin be deprecated? Are there any users of it that would object?
Andy -----Original Message----- From: Kleber, Ulrich (NSN - DE/Munich) [mailto:[email protected]] Sent: Thursday, March 04, 2010 8:48 AM To: [email protected]; [email protected] Subject: Re: [Openhpi-devel] openHPI plugins Hi, thanks a lot, this explains it. Do you think the ipmi plugin could be deprecated? Cheers, Uli > -----Original Message----- > From: ext Andy Cress [mailto:[email protected]] > Sent: Thursday, March 04, 2010 2:43 PM > To: [email protected]; [email protected] > Subject: Re: [Openhpi-devel] openHPI plugins > > RE: difference between the ipmi and ipmidirect plugins > In the beginning, the two plugins were targeted at different > segments of > IPMI servers (conventional and bladed), but now they are simply two > different approaches to the same goal. > > The ipmi plugin uses/requires OpenIPMI libraries in addition > to openhpi > in order to talk to the OpenIPMI driver. It has become > rather stale and > didn't work at all for Intel IPMI servers the last time I tested with > it. (bug 1565999 from 2006 is still Open) > > The ipmidirect plugin talks directly to the OpenIPMI driver, and works > fine for both conventional and bladed IPMI servers. This is > the choice > that we use. > > Andy > > -----Original Message----- > From: Kleber, Ulrich (NSN - DE/Munich) [mailto:[email protected]] > Sent: Thursday, March 04, 2010 3:33 AM > To: [email protected]; [email protected] > Subject: Re: [Openhpi-devel] openHPI > > Hi Lars, > I didn't see any reply to your email on the reflector, but I was > interested in the topic. I am not yet really an expert (yet) on > the plugin, but maybe together we can progress with your topics. > See inline. > Cheers, > Uli > > > -----Original Message----- > > From: ext Lars Wetzel [mailto:[email protected]] > > Sent: Tuesday, February 09, 2010 11:24 AM > > To: [email protected]; [email protected] > > Subject: Re: [Openhpi-devel] openHPI > > > > Hi Ric, > > > > yes, we debug the ipmi events as you describe below last week. > > We could see that the Mx->M0->Mx events are missing. So I > > think it isn't a > > problem of the openhpi/ipmidirect plugin. > > > > But I want to take the opportunity and ask some short > > questions regarding the > > ipmidirect plugin before you leave the project. Maybe you can > > help me to have > > a better understanding of the ipmidirect plugin background. I > > hope, I'm not > > too late. > > > > I know the openhpid/ipmidirect combination only by the code. > > I never run it in > > a system. > > - I think the openhpid/ipmidirect isn't programmed to replace > > a Shelf or > > ChassisManager in an xTCA system. Is this correct? > > I think this is true. > As far as I know, the ipmidirect plugin talks to the ChassisManager. > At least it worked when I configured the daemon that way. > However I am still a bit confused about the difference between > ipmi plugin and ipmidirect plugin. > > > > > > - I also miss some stuff from the xTCA Mapping Specification > > (e.g. a SLOT > > resource, like SYSTEM_CHASSIS - XYZ_SLOT - XYZ_RESOURCE). > > Should the plugin > > be SAF mapping specification compliant? > > I think the plugin should be compliant, but to which mapping spec? > It looks like you refer to the xTCA mapping spec, which is not > published yet. > As soon as the new mapping spec is published, we should start > working on a plugin compliant to that one. > > Hope that helps, > Cheers, > Uli > > > > > > > Thanks in forward and best wishes for the new job! > > Lars > > > > On Tuesday, 9. February 2010 02:47, Ric White wrote: > > > Hello Ayman, > > > > > > We tried to make the IPMI plug-ins as general purpose as > > possible, but > > > sometimes a bit of modification is required to make them > > play well with > > > specific hardware. > > > > > > To check if the daemon is receiving and processing the IMPI > > events for > > > hot swap as Lars suggested, you can add the following > parameters to > > > the libipmidirect handler stanza in your openhpi.conf file: > > > > > > logflags = "file" # "" means logging off; also use > > "file stdout" > > > logfile = "ipmidirect" # log file name prefix; ${logfile}.log > > > logfile_max = "10000" # maximum log file size in kilobytes > > > > > > This will create a ipmidirect.log file that could be used > > to see what is > > > really going on. > > > > > > On Tue, 2010-01-26 at 14:35 +0000, Ayman Daoud wrote: > > > > Dear openHPI representative, > > > > > > > > I have been working on a project to monitor the uTCA > > hardware using > > > > openHPI. I am using openhpi-2.14.1 with the ipmidirect > > plugin. During > > > > my work (using hpi_shell) I experienced the following > questionable > > > > behaviours which might be bugs: > > > > > > > > 1. if FRU is added to the chassis after the daemon has started, > > > > openHPI do not detect that FRU; No RPT entry added in the > > RPT table > > > > for the newly added FRU nor an event is generated to > indicate the > > > > addition of the FRU. (this is different from extracting FRU and > > > > reinstalling it which is fine except for what is stated > > in #2 and 3) > > > > > > > > 2. SAHPI_HS_STATE_NOT_PRESENT event is not generated when > > the FRU is > > > > removed from the chassis. > > > > > > > > 3. when FRU is removed from the chassis, the > > corresponding RPT entry > > > > is not deleted from the RPT table. > > > > > > > > 4. if the daemon start with a FRU plugged into the > chassis but the > > > > latch is not pushed in; we see a RPT entry for the > > resource modelling > > > > the FRU, but when the latch is pushed in, no event is > generated to > > > > indicate the transition from INACTIVE (or INSERTION > > PENDING) state to > > > > ACTIVE state. > > > > > > > > 5. saHpiHotSwapStateGet() return an error when it is called for > > > > resources that have the FRU capability but not the HS > > capability. the > > > > HPI specs states that this function should be enabled for > > resources > > > > with the FRU capability. > > > > > > This (your #5) appears to be a defect in the daemon. It is > > checking the > > > resource's ResourceCapabilities flag, and if > > > SAHPI_CAPABILITY_MANAGED_HOTSWAP is not set, it will always return > > > SA_ERR_HPI_CAPABILITY. According to the B.03.01 > > Specification, it should > > > instead be checking that SAHPI_CAPABILITY_FRU is set. Looks > > like this > > > was a change in behavior between the B.02.01 and B.03.01 HPI > > > Specifications. > > > > > > I have submitted bug #2948127 for this. > > > > > > Best Regards, > > > Ric White > > > > > > > Any help with these issues will be greatly appreciated. > > > > > > > > Best Regards, > > > > > > > > Ayman Doaud > > > > Software Engineer > > > > > > > > Tecore Networks > > > > > > > > Phone: +1 410.872.6286 > > > > Fax: +1 410.872.6010 > > > > e-mail: [email protected] > > > > > > > > > > > > THIS E-MAIL MAY CONTAIN PRIVILEGED, CONFIDENTIAL, > > COPYRIGHTED OR OTHER > > > > LEGALLY PROTECTED INFORMATION, AND IS INTENDED > EXCLUSIVELY FOR THE > > > > INTENDED RECIPIENT. IF YOU ARE NOT THE INTENDED RECIPIENT > > (EVEN IF THE > > > > E-MAIL ADDRESS ABOVE IS YOURS), YOU MAY NOT REVIEW, > > STORE, USE, COPY, > > > > DISCLOSE OR RETRANSMIT IT IN ANY FORM. IF YOU ARE NOT > THE INTENDED > > > > RECIPIENT OR OTHERWISE HAVE RECEIVED THIS BY MISTAKE, OR > > IF YOU WISH > > > > TO BE REMOVED FROM A MAILING LIST, PLEASE IMMEDIATELY NOTIFY THE > > > > SENDER BY RETURN E-MAIL (AND TECORE AT > [email protected]), THEN > > > > DELETE THE MESSAGE IN ITS ENTIRETY. THANK YOU. > > > > > > > > > > -------------------------------------------------------------- > > ----------- > > > >----- The Planet: dedicated and managed hosting, cloud > > storage, colocation > > > > Stay online with enterprise data centers and the best > > network in the > > > > business > > > > Choose flexible plans and management services without long-term > > > > contracts > > > > Personal 24x7 support from experience hosting pros just a > > phone call > > > > away. > > > > http://p.sf.net/sfu/theplanet-com > > > > _______________________________________________ > > > > Openhpi-devel mailing list > > > > [email protected] > > > > https://lists.sourceforge.net/lists/listinfo/openhpi-devel > > > > > > > > -------------------------------------------------------------- > > ------------- > > >--- The Planet: dedicated and managed hosting, cloud > > storage, colocation > > > Stay online with enterprise data centers and the best > network in the > > > business Choose flexible plans and management services > > without long-term > > > contracts Personal 24x7 support from experience hosting > > pros just a phone > > > call away. http://p.sf.net/sfu/theplanet-com > > > _______________________________________________ > > > Openhpi-devel mailing list > > > [email protected] > > > https://lists.sourceforge.net/lists/listinfo/openhpi-devel > > > > -- > > ------------------------------- > > Dipl. Wi.ing. > > Lars Wetzel > > Uttinger Str. 13 > > 86938 Schondorf a. Ammersee > > > > Tel.: 0179-2096845 > > Mail: [email protected] > > > > USt-IdNr.: DE181396006 > > > > -------------------------------------------------------------- > > ---------------- > > The Planet: dedicated and managed hosting, cloud storage, colocation > > Stay online with enterprise data centers and the best network > > in the business > > Choose flexible plans and management services without > > long-term contracts > > Personal 24x7 support from experience hosting pros just a > > phone call away. > > http://p.sf.net/sfu/theplanet-com > > _______________________________________________ > > Openhpi-devel mailing list > > [email protected] > > https://lists.sourceforge.net/lists/listinfo/openhpi-devel > > > > -------------------------------------------------------------- > ---------- > ------ > Download Intel® Parallel Studio Eval > Try the new software tools for yourself. Speed compiling, find bugs > proactively, and fine-tune applications for parallel performance. > See why Intel Parallel Studio got high marks during beta. > http://p.sf.net/sfu/intel-sw-dev > _______________________________________________ > Openhpi-devel mailing list > [email protected] > https://lists.sourceforge.net/lists/listinfo/openhpi-devel > ceforge.net/lists/listinfo/openhpi-devel > > > > -------------------------------------------------------------- > ---------------- > Download Intel® Parallel Studio Eval > Try the new software tools for yourself. Speed compiling, find bugs > proactively, and fine-tune applications for parallel performance. > See why Intel Parallel Studio got high marks during beta. > http://p.sf.net/sfu/intel-sw-dev > _______________________________________________ > Openhpi-devel mailing list > [email protected] > https://lists.sourceforge.net/lists/listinfo/openhpi-devel > ------------------------------------------------------------------------ ------ Download Intel® Parallel Studio Eval Try the new software tools for yourself. Speed compiling, find bugs proactively, and fine-tune applications for parallel performance. See why Intel Parallel Studio got high marks during beta. http://p.sf.net/sfu/intel-sw-dev _______________________________________________ Openhpi-devel mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/openhpi-devel ceforge.net/lists/listinfo/openhpi-devel ------------------------------------------------------------------------------ Download Intel® Parallel Studio Eval Try the new software tools for yourself. Speed compiling, find bugs proactively, and fine-tune applications for parallel performance. See why Intel Parallel Studio got high marks during beta. http://p.sf.net/sfu/intel-sw-dev _______________________________________________ Openhpi-devel mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/openhpi-devel
