Hi Everyone,
I have to apologize in advance. I have used other solutions before in regards 
to monitoring servers and having redundancy, but this is my first experience 
with heartbeat.
I've inherited a set of servers from someone that is using drbd to provide a 
redundant installation of mysql and another set of disks for lucene. After 
installing some monitoring software I started receiving these messages in 
emails:
Apr  4 22:14:15 staircase pengine: [31090]: notice: unpack_rsc_op: Operation 
lucene-disk:0_last_failure_0 found resource lucene-disk:0 active on 
staircase.bup.prod.localApr  4 22:14:15 staircase pengine: [31090]: notice: 
unpack_rsc_op: Operation db-master-mysql_last_failure_0 found resource 
db-master-mysql active on drawers.bup.prod.localApr  4 22:14:15 staircase 
pengine: [31090]: notice: unpack_rsc_op: Operation lucene-server_last_failure_0 
found resource lucene-server active on drawers.bup.prod.localApr  4 22:14:15 
staircase pengine: [31090]: notice: unpack_rsc_op: Operation 
lucene-disk:1_last_failure_0 found resource lucene-disk:1 active in master mode 
on drawers.bup.prod.localApr  4 22:14:15 staircase pengine: [31090]: notice: 
unpack_rsc_op: Operation lucene-fs_last_failure_0 found resource lucene-fs 
active on drawers.bup.prod.localApr  4 22:14:15 staircase pengine: [31090]: 
notice: unpack_rsc_op: Operation lucene-ip_last_failure_0 found resource 
lucene-ip active on drawers.bup.p
 rod.localApr  4 22:14:15 staircase pengine: [31090]: notice: LogActions: Leave 
  db-master-ip#011(Started drawers.bup.prod.local)Apr  4 22:14:15 staircase 
pengine: [31090]: notice: LogActions: Leave   db-master-mysql#011(Started 
drawers.bup.prod.local)Apr  4 22:14:15 staircase pengine: [31090]: notice: 
LogActions: Leave   db-slave-ip#011(Started staircase.bup.prod.local)Apr  4 
22:14:15 staircase pengine: [31090]: notice: LogActions: Leave   
db-slave-mysql#011(Started staircase.bup.prod.local))Apr  4 22:14:15 staircase 
pengine: [31090]: notice: LogActions: Leave   lucene-disk:0#011(Slave 
staircase.bup.prod.local)Apr  4 22:14:15 staircase pengine: [31090]: notice: 
LogActions: Leave   lucene-disk:1#011(Master drawers.bup.prod.local)Apr  4 
22:14:15 staircase pengine: [31090]: notice: LogActions: Leave   
lucene-fs#011(Started drawers.bup.prod.local)Apr  4 22:14:15 staircase pengine: 
[31090]: notice: LogActions: Leave   lucene-ip#011(Started 
drawers.bup.prod.local)Apr  4 22:14:15 staircas
 e pengine: [31090]: notice: LogActions: Leave   lucene-server#011(Started 
drawers.bup.prod.local)Apr  4 22:14:15 staircase pengine: [31090]: notice: 
process_pe_message: Transition 17386: PEngine Input stored in: 
/var/lib/pengine/pe-input-478.bz2Apr  4 22:14:15 staircase crmd: [31085]: info: 
do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ 
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]Apr  4 22:14:15 
staircase crmd: [31085]: info: unpack_graph: Unpacked transition 17386: 0 
actions in 0 synapsesApr  4 22:14:15 staircase crmd: [31085]: info: 
do_te_invoke: Processing graph 17386 (ref=pe_calc-dc-1396664055-18368) derived 
from /var/lib/pengine/pe-input-478.bz2Apr  4 22:14:15 staircase crmd: [31085]: 
info: run_graph: ====================================================Apr  4 
22:14:15 staircase crmd: [31085]: notice: run_graph: Transition 17386 
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pengine/pe-input-478.bz2): 
 CompleteApr  4 22:14:15 staircase crmd: [31085]: info: te_graph_trigger: 
Transition 17386 is now completeApr  4 22:14:15 staircase crmd: [31085]: info: 
notify_crmd: Transition 17386 status: done - <null>Apr  4 22:14:15 staircase 
crmd: [31085]: info: do_state_transition: State transition S_TRANSITION_ENGINE 
-> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]Apr  4 
22:14:15 staircase crmd: [31085]: info: do_state_transition: Starting PEngine 
Recheck TimerApr  4 22:14:22 staircase lrmd: [31082]: info: rsc:lucene-disk:0 
monitor[286] (pid 12190)Apr  4 22:14:22 staircase lrmd: [31082]: info: 
operation monitor[286] on lucene-disk:0 for client 31085: pid 12190 exited with 
return code 0
Everything seems fine, and the services are now working after I found some 
issues with the daemon on one box, fixed it, then cleaned and reprobed with the 
crmadmin command. I am not sure if anything is actually wrong, or if these are 
harmless warnings. Is it anything I should worry about? If so, I can provide my 
config files. I just wasn't sure if this is some sort of harmless thing or if I 
should genuinely be worried.
Thanks for your help!Kathy
                                          
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to