RE: difference between the ipmi and ipmidirect plugins
In the beginning, the two plugins were targeted at different segments of
IPMI servers (conventional and bladed), but now they are simply two
different approaches to the same goal.  

The ipmi plugin uses/requires OpenIPMI libraries in addition to openhpi
in order to talk to the OpenIPMI driver.  It has become rather stale and
didn't work at all for Intel IPMI servers the last time I tested with
it.  (bug 1565999 from 2006 is still Open)

The ipmidirect plugin talks directly to the OpenIPMI driver, and works
fine for both conventional and bladed IPMI servers.  This is the choice
that we use.   

Andy

-----Original Message-----
From: Kleber, Ulrich (NSN - DE/Munich) [mailto:[email protected]] 
Sent: Thursday, March 04, 2010 3:33 AM
To: [email protected]; [email protected]
Subject: Re: [Openhpi-devel] openHPI 

Hi Lars,
I didn't see any reply to your email on the reflector, but I was
interested in the topic. I am not yet really an expert (yet) on 
the plugin, but maybe together we can progress with your topics.
See inline.
Cheers,
Uli 

> -----Original Message-----
> From: ext Lars Wetzel [mailto:[email protected]] 
> Sent: Tuesday, February 09, 2010 11:24 AM
> To: [email protected]; [email protected]
> Subject: Re: [Openhpi-devel] openHPI
> 
> Hi Ric,
> 
> yes, we debug the ipmi events as you describe below last week.
> We could see that the Mx->M0->Mx events are missing.  So I 
> think it isn't a 
> problem of the openhpi/ipmidirect plugin.
> 
> But I want to take the opportunity and ask some short 
> questions regarding the 
> ipmidirect plugin before you leave the project. Maybe you can 
> help me to have 
> a better understanding of the ipmidirect plugin background. I 
> hope, I'm not 
> too late.
> 
> I know the openhpid/ipmidirect combination only by the code. 
> I never run it in 
> a system. 
> - I think the openhpid/ipmidirect isn't programmed to replace 
> a Shelf or 
> ChassisManager in an xTCA system. Is this correct?

I think this is true.
As far as I know, the ipmidirect plugin talks to the ChassisManager.
At least it worked when I configured the daemon that way.
However I am still a bit confused about the difference between
ipmi plugin and ipmidirect plugin.


> 
> - I also miss some stuff from the xTCA Mapping Specification 
> (e.g. a SLOT 
> resource, like SYSTEM_CHASSIS - XYZ_SLOT - XYZ_RESOURCE). 
> Should the plugin  
> be SAF mapping specification compliant?

I think the plugin should be compliant, but to which mapping spec?
It looks like you refer to the xTCA mapping spec, which is not
published yet.
As soon as the new mapping spec is published, we should start
working on a plugin compliant to that one.

Hope that helps,
Cheers,
Uli



> 
> Thanks in forward and best wishes for the new job!
> Lars
> 
> On Tuesday, 9. February 2010 02:47, Ric White wrote:
> > Hello Ayman,
> >
> > We tried to make the IPMI plug-ins as general purpose as 
> possible, but
> > sometimes a bit of modification is  required to make them 
> play well with
> > specific hardware.
> >
> > To check if the daemon is receiving and processing the IMPI 
> events for
> > hot swap as Lars suggested, you can add the following parameters to
> > the libipmidirect handler stanza in your openhpi.conf file:
> >
> >     logflags = "file"    # "" means logging off; also use 
> "file stdout"
> >     logfile = "ipmidirect"  # log file name prefix; ${logfile}.log
> >     logfile_max = "10000" # maximum log file size in kilobytes
> >
> > This will create a ipmidirect.log file that could be used 
> to see what is
> > really going on.
> >
> > On Tue, 2010-01-26 at 14:35 +0000, Ayman Daoud wrote:
> > > Dear openHPI representative,
> > >
> > > I have been working on a project to monitor the uTCA 
> hardware using
> > > openHPI. I am using openhpi-2.14.1 with the ipmidirect 
> plugin. During
> > > my work (using hpi_shell) I experienced the following questionable
> > > behaviours which might be bugs:
> > >
> > > 1. if FRU is added to the chassis after the daemon has started,
> > > openHPI do not detect that FRU; No RPT entry added in the 
> RPT table
> > > for the newly added FRU nor an event is generated to indicate the
> > > addition of the FRU. (this is different from extracting FRU and
> > > reinstalling it which is fine except for what is stated 
> in #2 and 3)
> > >
> > > 2. SAHPI_HS_STATE_NOT_PRESENT event is not generated when 
> the FRU is
> > > removed from the chassis.
> > >
> > > 3. when FRU is removed from the chassis, the 
> corresponding RPT entry
> > > is not deleted from the RPT table.
> > >
> > > 4. if the daemon start with a FRU plugged into the chassis but the
> > > latch is not pushed in; we see a RPT entry for the 
> resource modelling
> > > the FRU, but when the latch is pushed in, no event is generated to
> > > indicate the transition from INACTIVE (or INSERTION 
> PENDING) state to
> > > ACTIVE state.
> > >
> > > 5. saHpiHotSwapStateGet() return an error when it is called for
> > > resources that have the FRU capability but not the HS 
> capability. the
> > > HPI specs states that this function should be enabled for 
> resources
> > > with the FRU capability.
> >
> > This (your #5) appears to be a defect in the daemon. It is 
> checking the
> > resource's ResourceCapabilities flag, and if
> > SAHPI_CAPABILITY_MANAGED_HOTSWAP is not set, it will always return
> > SA_ERR_HPI_CAPABILITY. According to the B.03.01 
> Specification, it should
> > instead be checking that SAHPI_CAPABILITY_FRU is set. Looks 
> like this
> > was a change in behavior between the B.02.01 and B.03.01 HPI
> > Specifications.
> >
> > I have submitted bug #2948127 for this.
> >
> > Best Regards,
> > Ric White
> >
> > > Any help with these issues will be greatly appreciated.
> > >
> > > Best Regards,
> > >
> > > Ayman Doaud
> > > Software Engineer
> > >
> > > Tecore Networks
> > >
> > > Phone: +1 410.872.6286
> > > Fax: +1 410.872.6010
> > > e-mail: [email protected]
> > >
> > >
> > > THIS E-MAIL MAY CONTAIN PRIVILEGED, CONFIDENTIAL, 
> COPYRIGHTED OR OTHER
> > > LEGALLY PROTECTED INFORMATION, AND IS INTENDED EXCLUSIVELY FOR THE
> > > INTENDED RECIPIENT. IF YOU ARE NOT THE INTENDED RECIPIENT 
> (EVEN IF THE
> > > E-MAIL ADDRESS ABOVE IS YOURS), YOU MAY NOT REVIEW, 
> STORE, USE, COPY,
> > > DISCLOSE OR RETRANSMIT IT IN ANY FORM. IF YOU ARE NOT THE INTENDED
> > > RECIPIENT OR OTHERWISE HAVE RECEIVED THIS BY MISTAKE, OR 
> IF YOU WISH
> > > TO BE REMOVED FROM A MAILING LIST, PLEASE IMMEDIATELY NOTIFY THE
> > > SENDER BY RETURN E-MAIL (AND TECORE AT [email protected]), THEN
> > > DELETE THE MESSAGE IN ITS ENTIRETY. THANK YOU.
> > >
> > > 
> --------------------------------------------------------------
> -----------
> > >----- The Planet: dedicated and managed hosting, cloud 
> storage, colocation
> > > Stay online with enterprise data centers and the best 
> network in the
> > > business
> > > Choose flexible plans and management services without long-term
> > > contracts
> > > Personal 24x7 support from experience hosting pros just a 
> phone call
> > > away.
> > > http://p.sf.net/sfu/theplanet-com
> > > _______________________________________________
> > > Openhpi-devel mailing list
> > > [email protected]
> > > https://lists.sourceforge.net/lists/listinfo/openhpi-devel
> >
> > 
> --------------------------------------------------------------
> -------------
> >--- The Planet: dedicated and managed hosting, cloud 
> storage, colocation
> > Stay online with enterprise data centers and the best network in the
> > business Choose flexible plans and management services 
> without long-term
> > contracts Personal 24x7 support from experience hosting 
> pros just a phone
> > call away. http://p.sf.net/sfu/theplanet-com
> > _______________________________________________
> > Openhpi-devel mailing list
> > [email protected]
> > https://lists.sourceforge.net/lists/listinfo/openhpi-devel
> 
> -- 
> -------------------------------
> Dipl. Wi.ing.
> Lars Wetzel
> Uttinger Str. 13
> 86938 Schondorf a. Ammersee
> 
> Tel.: 0179-2096845
> Mail: [email protected]
> 
> USt-IdNr.: DE181396006
> 
> --------------------------------------------------------------
> ----------------
> The Planet: dedicated and managed hosting, cloud storage, colocation
> Stay online with enterprise data centers and the best network 
> in the business
> Choose flexible plans and management services without 
> long-term contracts
> Personal 24x7 support from experience hosting pros just a 
> phone call away.
> http://p.sf.net/sfu/theplanet-com
> _______________________________________________
> Openhpi-devel mailing list
> [email protected]
> https://lists.sourceforge.net/lists/listinfo/openhpi-devel
> 

------------------------------------------------------------------------
------
Download Intel® Parallel Studio Eval
Try the new software tools for yourself. Speed compiling, find bugs
proactively, and fine-tune applications for parallel performance.
See why Intel Parallel Studio got high marks during beta.
http://p.sf.net/sfu/intel-sw-dev
_______________________________________________
Openhpi-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/openhpi-devel
ceforge.net/lists/listinfo/openhpi-devel



------------------------------------------------------------------------------
Download Intel® Parallel Studio Eval
Try the new software tools for yourself. Speed compiling, find bugs
proactively, and fine-tune applications for parallel performance.
See why Intel Parallel Studio got high marks during beta.
http://p.sf.net/sfu/intel-sw-dev
_______________________________________________
Openhpi-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/openhpi-devel

Reply via email to