Hi list,

I use net-snmp 5.4.2.1 (on SLES 11 SP1) with agentx to monitor GPFS 
(v3.4) distributed filesystems.

This works by running a gpfs snmp agentx daemon on one of the gpfs nodes.

Usually, I can snmpwalk starting at oid gpfsDiskStatus and get a correct 
list of the status of all disks on all nodes.

But for one node the agent delivers wrong results after disks of one 
node have failed and got repaired.

(I only show one disk per node. The other disks show the same behaviour)

GPFS-MIB::gpfsDiskStatus."data"."down"."data06node19" = STRING: "unknown"
GPFS-MIB::gpfsDiskStatus."data"."hddpool"."data06node17" = STRING: "InUse"
GPFS-MIB::gpfsDiskStatus."data"."hddpool"."data06node18" = STRING: "InUse"

As you can see, the oid of the repaired node contains "down" instead of 
the disk pool name.

I verified with strace the gpfs snmp agentx daemon (on node17) recieves 
the correct status "InUse" for node19 disks.
I guess, it cannot deliver that becaues the oid is wrong (contains 
"down" instead of "hddpool").
Of course I restarted all gpfs daemons, the gpfs agentx daemon, snmpd: 
no change.
On other installations I already had disk problems and repaired them. 
The snmp disk status then was always correct.

I couldnt find any files that contain the wrong oids - not on the snmpd 
side and not on the gpfs side.
So my question is: How does the oid registration work with the agentx 
protocol?
Is there some caching of oids involved that I overlooked?
Maybe someone had a similar problem and solved this already?

BR,
Joachim


------------------------------------------------------------------------------
Everyone hates slow websites. So do we.
Make your web apps faster with AppDynamics
Download AppDynamics Lite for free today:
http://p.sf.net/sfu/appdyn_d2d_feb
_______________________________________________
Net-snmp-users mailing list
Net-snmp-users@lists.sourceforge.net
Please see the following page to unsubscribe or change other options:
https://lists.sourceforge.net/lists/listinfo/net-snmp-users

Reply via email to