Hi All,
I had a few extra minutes and I upgraded to heartbeat to version 2.1.3
and restarted both of my servers.
I have things working on just fine on the "master" server, but I have
several issues..
1) drbd seems to not be communicating..
on the master
thebrain:~ # cat /proc/drbd
version: 8.0.6 (api:86/proto:86)
SVN Revision: 3048 build by [EMAIL PROTECTED], 2007-09-03 10:39:27
0: cs:StandAlone st:Primary/Unknown ds:UpToDate/DUnknown r---
ns:0 nr:0 dw:10180 dr:39945 al:488 bm:488 lo:0 pe:0 ua:0 ap:0
resync: used:0/31 hits:0 misses:0 starving:0 dirty:0 changed:0
act_log: used:0/127 hits:2057 misses:505 starving:15 dirty:2 changed:
488
on the slave
pinky:~ # cat /proc/drbd
version: 8.0.6 (api:86/proto:86)
SVN Revision: 3048 build by [EMAIL PROTECTED], 2007-09-03 10:39:27
0: cs:StandAlone st:Secondary/Unknown ds:UpToDate/DUnknown r---
ns:0 nr:0 dw:0 dr:0 al:0 bm:106 lo:0 pe:0 ua:0 ap:0
resync: used:0/31 hits:0 misses:0 starving:0 dirty:0 changed:0
act_log: used:0/127 hits:0 misses:0 starving:0 dirty:0 changed:0
looking at the logs on the client .. I find
pinky:~ # grep drbd /var/log/messages | tail -20
Jan 31 01:12:33 pinky crmd: [4080]: info: process_lrm_event: LRM
operation drbd_sys:1_notify_0 (call=18, rc=0) complete
Jan 31 01:12:34 pinky crmd: [4080]: info: do_lrm_rsc_op: Performing
op=drbd_sys:1_notify_0 key=75:20:ab1e279e-4466-4cee-a6b7-47c81743e85d)
Jan 31 01:12:34 pinky lrmd: [4077]: info: rsc:drbd_sys:1: notify
Jan 31 01:12:34 pinky drbd[4761]: [4772]: DEBUG: mirror notify: pre
for promote - counts: active 0 - starting 2 - stopping 0
Jan 31 01:12:34 pinky crmd: [4080]: info: process_lrm_event: LRM
operation drbd_sys:1_notify_0 (call=19, rc=0) complete
Jan 31 01:12:35 pinky crmd: [4080]: info: do_lrm_rsc_op: Performing
op=drbd_sys:1_notify_0 key=76:20:ab1e279e-4466-4cee-a6b7-47c81743e85d)
Jan 31 01:12:35 pinky lrmd: [4077]: info: rsc:drbd_sys:1: notify
Jan 31 01:12:35 pinky drbd[4773]: [4784]: DEBUG: mirror notify: post
for promote - counts: active 0 - starting 2 - stopping 0
Jan 31 01:12:35 pinky drbd[4773]: [4786]: DEBUG: mirror: Calling
drbdadm -c /etc/drbd.conf state mirror
Jan 31 01:12:35 pinky drbd[4773]: [4793]: DEBUG: mirror: Exit code 0
Jan 31 01:12:35 pinky drbd[4773]: [4794]: DEBUG: mirror: Command
output: Secondary/Unknown
Jan 31 01:12:35 pinky drbd[4773]: [4802]: DEBUG: mirror: Calling
drbdadm -c /etc/drbd.conf cstate mirror
Jan 31 01:12:35 pinky drbd[4773]: [4809]: DEBUG: mirror: Exit code 0
Jan 31 01:12:35 pinky drbd[4773]: [4810]: DEBUG: mirror: Command
output: StandAlone
Jan 31 01:12:35 pinky drbd[4773]: [4811]: DEBUG: mirror status:
Secondary/Unknown Secondary Unknown StandAlone
Jan 31 01:12:35 pinky drbd[4773]: [4812]: DEBUG: mirror: Calling /usr/
sbin/crm_master -v 5 -l reboot
Jan 31 01:12:37 pinky drbd[4773]: [4817]: DEBUG: mirror: Exit code 0
Jan 31 01:12:37 pinky drbd[4773]: [4818]: DEBUG: mirror: Command
output: No set matching id=master-6deaf16f-a98f-4c56-af00-a97be99f2e68
in status
Jan 31 01:12:37 pinky lrmd: [4077]: info: RA output: (drbd_sys:
1:notify:stdout) No set matching id=master-6deaf16f-a98f-4c56-af00-
a97be99f2e68 in status
Jan 31 01:12:37 pinky crmd: [4080]: info: process_lrm_event: LRM
operation drbd_sys:1_notify_0 (call=20, rc=0) complete
I'm tired and google has failed me, I don't know where to start
looking for what's wrong.
The other problem I noticed, is that mgmtd is no longer starting
Jan 31 01:12:06 pinky heartbeat: [3821]: info: Starting child client "/
usr/lib64/heartbeat/mgmtd -v" (0,0)
Jan 31 01:12:06 pinky heartbeat: [4081]: info: Starting "/usr/lib64/
heartbeat/mgmtd -v" as uid 0 gid 0 (pid 4081)
Jan 31 01:12:06 pinky mgmtd: [4081]: info: G_main_add_SignalHandler:
Added signal handler for signal 15
Jan 31 01:12:06 pinky mgmtd: [4081]: debug: Enabling coredumps
Jan 31 01:12:06 pinky mgmtd: [4081]: info: G_main_add_SignalHandler:
Added signal handler for signal 10
Jan 31 01:12:06 pinky mgmtd: [4081]: info: G_main_add_SignalHandler:
Added signal handler for signal 12
Jan 31 01:12:06 pinky heartbeat: [3821]: WARN: Client [mgmtd] pid 4081
failed authorization [no default client auth]
Jan 31 01:12:06 pinky heartbeat: [3821]: ERROR:
api_process_registration_msg: cannot add client(mgmtd)
Jan 31 01:12:06 pinky mgmtd: [4081]: ERROR: Cannot sign on with
heartbeat
Jan 31 01:12:06 pinky mgmtd: [4081]: ERROR: REASON:
Jan 31 01:12:06 pinky mgmtd: [4081]: ERROR: Can't initialize
management library.Shutting down.(-1)
Jan 31 01:12:06 pinky heartbeat: [3821]: WARN: Managed /usr/lib64/
heartbeat/mgmtd -v process 4081 exited with return code 1.
Jan 31 01:12:06 pinky heartbeat: [3821]: ERROR: Respawning client "/
usr/lib64/heartbeat/mgmtd -v":
Jan 31 01:12:06 pinky heartbeat: [3821]: info: Starting child client "/
usr/lib64/heartbeat/mgmtd -v" (0,0)
Jan 31 01:12:07 pinky heartbeat: [4084]: info: Starting "/usr/lib64/
heartbeat/mgmtd -v" as uid 0 gid 0 (pid 4084)
Jan 31 01:12:07 pinky mgmtd: [4084]: info: G_main_add_SignalHandler:
Added signal handler for signal 15
Jan 31 01:12:07 pinky mgmtd: [4084]: debug: Enabling coredumps
Jan 31 01:12:07 pinky mgmtd: [4084]: info: G_main_add_SignalHandler:
Added signal handler for signal 10
Jan 31 01:12:07 pinky mgmtd: [4084]: info: G_main_add_SignalHandler:
Added signal handler for signal 12
Jan 31 01:12:07 pinky heartbeat: [3821]: WARN: Client [mgmtd] pid 4084
failed authorization [no default client auth]
Jan 31 01:12:07 pinky heartbeat: [3821]: ERROR:
api_process_registration_msg: cannot add client(mgmtd)
Jan 31 01:12:07 pinky mgmtd: [4084]: ERROR: Cannot sign on with
heartbeat
Jan 31 01:12:07 pinky mgmtd: [4084]: ERROR: REASON:
Jan 31 01:12:07 pinky mgmtd: [4084]: ERROR: Can't initialize
management library.Shutting down.(-1)
Jan 31 01:12:07 pinky heartbeat: [3821]: WARN: Managed /usr/lib64/
heartbeat/mgmtd -v process 4084 exited with return code 1.
Jan 31 01:12:07 pinky heartbeat: [3821]: ERROR: Respawning client "/
usr/lib64/heartbeat/mgmtd -v":
I see this on both the main and the secondary server.
I found some info at
http://www.gossamer-threads.com/lists/linuxha/users/45403?do=post_view_threaded#45403
but nothing that seemed to really apply..
If anyone can offer any suggestions, I would really appreciate it.
Thank you very much
Tony Nelson
Starpoint Solutions
On Jan 30, 2008, at 10:05 AM, Andreas Mock wrote:
-----Ursprüngliche Nachricht-----
Von: General Linux-HA mailing list <[email protected]>
Gesendet: 30.01.08 13:44:53
An: General Linux-HA mailing list <[email protected]>
co-incidentally, i've been thinking about such a feature recently...
i'm inclined to think that this functionality should be in the LRM
(ie. its a threshold for escalating to the CRM).
thoughts?
My thoughts if anyone is interested:
The result of the monitor action should be:
a) Resource is running.
b) Resource is not running.
But does it imply that it is running healthy? As the result of
the monitor action determines what happens to the resource,
I would say "YES". So the question is:
Does the resource run in a way able to fulfill it's service? (Yes/No)
But IMHO this question implies that the RA tries/should try to do as
much
as necessary to test the service-ability. This can be pretty much.
Sometimes too much if the service is doing what it should: Working
hard!
Timeout means: Nothing, no answer.
What shall I do with this information? As I said before: Shall I
assume
that something is wrong or everything is o.k.
Is everything o.k. if a service is producing so much load that I'm
not even
able to get the output of 'ps -ax'?
What is the difference between aksing once with long timeout and
multiple
times with short timeout. (If I measure the time, I could know in
both cases
that the RA need (too) long to get an answer).
IMHO not the timeout of the monitor action is the big problem, but
the possible
chain reaction you get after this:
monitor timeout because of heavy (regulsr) load => stop action
triggered =>
stop action times out (RAs try to do a graceful shutdown) because of
heavy load =>
MESS (node fencing or resource in unmanaged state).
My proposal: Make the timeout for monitor long enough. If timeout
occures assume
that there is really something wrong because a "simple" monitor
action does not work.
=> Stop resource. Implement a two-step stop action (probably in the
RA itself):
1) Try a graceful stop of resource. (e.d. db shutdown)
2) After inner timeout stop/kill resoure brutal (if possible)
3) If this doesn't work, signal timeout to upper instance which
results in known behaviour.
What would someone win: Kill one resource brutally with the hope all
other resources
still remain intact.
Of course interested to hear other aspects. :-)
Best regards
Andreas Mock
_____________________________________________________________________
Unbegrenzter Speicherplatz für Ihr E-Mail Postfach? Jetzt aktivieren!
http://www.digitaledienste.web.de/freemail/club/lp/?lp=7
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems