Hi there,

I am currently working on a 2 node drbd+pacemaker mysql active / passive 
cluster under SLES11 SP1 + HAE on VMware vSphere 4. The Clusters working quite 
well when I`m testing failover situations etc., but irregularly reboots both 
nodes without logging the reason for the reboots. As nothing like stonith 
shutdown commands or error notices appear in the system logs, it looks like the 
virtual machine got powered off suddenly. Other VMs (yet SLES10 with Heartbeat 
+ Stonith) running without any problems on the virtual infrastructure.. 
Yesterday I set up 2 clones of the virtual machines and configured them running 
without stonith device. These 2 clones survived the night. The ones I`m talking 
about right now shut down at different time (one at 7:57pm, one at 05:19 am 
this morning). Could be a coincidence but to be honest I don`t think so, 
because I got the same problems for weeks now, even after completely 
configuring the VMs from OS Installation until cluster configuration.. I will 
monitor the clones without stonith device over the weekend to see if theyre 
running without rebooting. Anyone got similar problems or can tell me where to 
look for restricting what could be the problem?

Thanks in advance!


In order to help you understand my configuration, some facts:

Summary of setup:
- 2 x SLES11 SP1 + HAE virtual machines on VMware vSphere4
- both VMs running on same ESX
- VMware Tools installed
- 2 x vNICs: 1 for Public Interface, 1 for Private Interconnect
- 3 vDisks: 1 for OS, 1 for DRBD, 1 shared disc for Stonith SBD
- pacemaker-1.1.2-0.2.1
- openais-1.1.2-0.5.19
- drbd-8.3.7-0.4.15
- Active / Passive Configuration


lsmod | grep softdog

softdog                 4055  2


/etc/sysconfig/sbd

SBD_DEVICE="/dev/sdc"
SBD_OPTS="-W"


/var/log/warn
...
...
Jan 21 09:50:09 mysql-node1 stonith: external/sbd device OK.
Jan 21 09:51:10 mysql-node1 stonith: external/sbd device OK.
Jan 21 09:52:11 mysql-node1 stonith: external/sbd device OK.
...
...


CRM Configure

node mysql-node1 \
        attributes standby="off"
node mysql-node2 \
        attributes standby="off"
primitive drbd_mysql ocf:linbit:drbd \
        params drbd_resource="mysql" \
        op monitor interval="15s"
primitive fs_mysql ocf:heartbeat:Filesystem \
        params device="/dev/drbd0" directory="/var/lib/mysql" fstype="ext3"
primitive ip_mysql ocf:heartbeat:IPaddr2 \
        params ip="192.168.96.10" nic="eth0"
primitive mysqld lsb:mysql
primitive notification ocf:heartbeat:MailTo \
        meta target-role="Started" \
        operations $id="notification-operations" \
        op monitor interval="10" timeout="10" \
        params email="[email protected]" subject="Cluster Takeover on mysql-cluster"
primitive ping ocf:pacemaker:ping \
        op monitor interval="15" timeout="5" \
        params host_list="192.168.96.2"
primitive stonith-device stonith:external/sbd \
        params sbd_device="/dev/sdc" \
        op monitor interval="60s"
group mysql fs_mysql ip_mysql mysqld
ms ms_drbd_mysql drbd_mysql \
        meta master-max="1" master-node-max="1" clone-max="2" 
clone-node-max="1" notify="true"
clone notificationclone notification
clone pingclone ping \
        meta globally-unique="false"
clone stonith stonith-device
location location_mysql mysql \
        rule $id="location_mysql-rule" -inf: not_defined pingd or pingd lte 0
colocation mysql_on_drbd inf: mysql ms_drbd_mysql:Master
order mysql_after_drbd inf: ms_drbd_mysql:promote mysql:start
property $id="cib-bootstrap-options" \
        dc-version="1.1.2-2e096a41a5f9e184a1c1537c82c6da1093698eb5" \
        cluster-infrastructure="openais" \
        expected-quorum-votes="2" \
        last-lrm-refresh="1294994548" \
        no-quorum-policy="ignore" \
        stonith-enabled="true"


Corosync.conf

compatibility: whitetank

aisexec {
        user: root
        group: root
}

totem {
        version: 2
        token: 3000
        join: 1000
        token_retransmits_before_loss_const: 20
        consensus: 3600
        max_messages: 20
        clear_node_high_bit: yes
        secauth: off
        threads: 0
        interface {
                ringnumber: 0
                bindnetaddr: 10.0.0.0
                mcastaddr: 226.94.1.1
                mcastport: 5405
        }
}

logging {
        fileline: off
        to_stderr: no
        to_logfile: no
        to_syslog: yes
        syslog_facility: daemon
        debug: on
        timestamp: on
}

amf {
        mode: disabled
}

service {
        name: pacemaker
        ver:  0
        use_mgmtd: yes
        use_logd: yes
}

/etc/drbd.d/mysql.res

resource mysql {
  protocol C;
  startup {
    wfc-timeout  0;
    degr-wfc-timeout 120;
  }
  disk {
    on-io-error   detach;
  }

  net {
    max-buffers     2048;
    ko-count 4;
    after-sb-0pri disconnect;
    after-sb-1pri disconnect;
    after-sb-2pri disconnect;
    rr-conflict disconnect;
  }
  syncer {
    rate 100M;
    al-extents 257;
  }
  on mysql-node1 {
    device     /dev/drbd0;
    disk       /dev/sdb;
    address    10.0.0.101:7788;
    meta-disk  internal;

  }

  on mysql-node2 {
    device    /dev/drbd0;
    disk      /dev/sdb;
    address   10.0.0.102:7788;
    meta-disk internal;
  }
}


/etc/drbd.d/global_common.conf

global {
        # usage-count yes;
        usage-count no;
        # minor-count dialog-refresh disable-ip-verification
}

common {
        protocol C;

        handlers {
                pri-on-incon-degr "/usr/lib/drbd/notify-pri-on-incon-degr.sh; 
/usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot 
-f";
                pri-lost-after-sb "/usr/lib/drbd/notify-pri-lost-after-sb.sh; 
/usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot 
-f";
                local-io-error "/usr/lib/drbd/notify-io-error.sh; 
/usr/lib/drbd/notify-emergency-shutdown.sh; echo o > /proc/sysrq-trigger ; halt 
-f";

        }

        startup {

        }

        disk {

        }

        net {

        }

        syncer {
                rate 100M;
        }
}




Node1 debug log:


Jan 22 10:24:17 mysql-node1 attrd_updater: [14240]: info: attrd_lazy_update: 
Connecting to cluster... 5 retries remaining

Jan 22 10:24:17 mysql-node1 attrd_updater: [14240]: debug: 
init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/attrd

Jan 22 10:24:17 mysql-node1 attrd: [9245]: debug: attrd_local_callback: update 
message from attrd_updater: pingd=1

Jan 22 10:24:17 mysql-node1 attrd: [9245]: debug: attrd_local_callback: 
Supplied: 1, Current: 1, Stored: 1

Jan 22 10:24:17 mysql-node1 attrd_updater: [14240]: debug: attrd_update: Sent 
update: pingd=1 for localhost

Jan 22 10:24:18 mysql-node1 lrmd: [9244]: debug: rsc:notification:0:41: monitor

Jan 22 10:24:18 mysql-node1 lrmd: [14241]: debug: perform_ra_op: resetting 
scheduler class to SCHED_OTHER

Jan 22 10:24:18 mysql-node1 lrmd: [9244]: debug: RA output: 
(notification:0:monitor:stdout) running

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: crm_timer_popped: PEngine 
Recheck Timer (I_PE_CALC) just popped!

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: s_crmd_fsa: Processing 
I_PE_CALC: [ state=S_IDLE cause=C_TIMER_POPPED origin=crm_timer_popped ]

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_state_transition: State 
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED 
origin=crm_timer_popped ]

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_state_transition: Progressed 
to state S_POLICY_ENGINE after C_TIMER_POPPED

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_state_transition: All 2 
cluster nodes are eligible to run resources.

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_DC_TIMER_STOP

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_INTEGRATE_TIMER_STOP

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_FINALIZE_TIMER_STOP

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_PE_INVOKE

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_pe_invoke: Query 207: 
Requesting the current CIB: S_POLICY_ENGINE

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_pe_invoke_callback: Invoking 
the PE: query=207, ref=pe_calc-dc-1295688261-214, seq=220, quorate=1

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: unpack_config: Startup 
probes: enabled

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: s_crmd_fsa: Processing 
I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE 
origin=handle_response ]

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: unpack_config: STONITH 
timeout: 60000

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_LOG

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: unpack_config: STONITH of 
failed nodes is enabled

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_state_transition: State 
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
cause=C_IPC_MESSAGE origin=handle_response ]

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: unpack_config: Stop all 
active resources: false

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_DC_TIMER_STOP

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: unpack_config: Cluster is 
symmetric - resources can run anywhere by default

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_INTEGRATE_TIMER_STOP

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: unpack_config: Default 
stickiness: 0

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_FINALIZE_TIMER_STOP

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: unpack_config: On loss of 
CCM Quorum: Ignore

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_TE_INVOKE

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: unpack_config: Node scores: 
'red' = -INFINITY, 'yellow' = 0, 'green' = 0

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: unpack_graph: Unpacked 
transition 114: 0 actions in 0 synapses

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: unpack_domains: Unpacking 
domains

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_te_invoke: Processing graph 
114 (ref=pe_calc-dc-1295688261-214) derived from 
/var/lib/pengine/pe-input-659.bz2

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: determine_online_status: 
Node mysql-node1 is online

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: print_graph: ## Empty 
transition graph ##

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: determine_online_status: 
Node mysql-node2 is online

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: run_graph: 
====================================================

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: find_clone: Internally 
renamed drbd_mysql:0 on mysql-node2 to drbd_mysql:1

Jan 22 10:24:21 mysql-node1 crmd: [9247]: notice: run_graph: Transition 114 
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pengine/pe-input-659.bz2): Complete

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: find_clone: Internally 
renamed stonith-device:0 on mysql-node2 to stonith-device:1

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: print_graph: ## Empty 
transition graph ##

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: find_clone: Internally 
renamed notification:0 on mysql-node2 to notification:1

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: te_graph_trigger: Transition 
114 is now complete

Jan 22 10:24:21 mysql-node1 pengine: [9246]: info: find_clone: Internally 
renamed ping:0 on mysql-node2 to ping:1

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: notify_crmd: Processing 
transition completion in state S_TRANSITION_ENGINE

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: clone_print:  Master/Slave 
Set: ms_drbd_mysql

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: notify_crmd: Transition 114 
status: done - <null>

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: native_active: Resource 
drbd_mysql:0 active on mysql-node1

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: s_crmd_fsa: Processing 
I_TE_SUCCESS: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL 
origin=notify_crmd ]

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: native_active: Resource 
drbd_mysql:0 active on mysql-node1

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_LOG

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: native_active: Resource 
drbd_mysql:1 active on mysql-node2

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_state_transition: State 
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS 
cause=C_FSA_INTERNAL origin=notify_crmd ]

Jan 22 10:24:21 mysql-node1 pengine: [9246]: debug: native_active: Resource 
drbd_mysql:1 active on mysql-node2

Jan 22 10:24:21 mysql-node1 crmd: [9247]: info: do_state_transition: Starting 
PEngine Recheck Timer

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: short_print:      Masters: 
[ mysql-node2 ]

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: crm_timer_start: Started 
PEngine Recheck Timer (I_PE_CALC:900000ms), src=400

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: short_print:      Slaves: 
[ mysql-node1 ]

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_DC_TIMER_STOP

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: group_print:  Resource 
Group: mysql

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_INTEGRATE_TIMER_STOP

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: native_print:      
fs_mysql        (ocf::heartbeat:Filesystem):    Started mysql-node2

Jan 22 10:24:21 mysql-node1 crmd: [9247]: debug: do_fsa_action: actions:trace:  
// A_FINALIZE_TIMER_STOP

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: native_print:      
ip_mysql        (ocf::heartbeat:IPaddr2):       Started mysql-node2

Jan 22 10:24:21 mysql-node1 pengine: [9246]: notice: native_print:      mysqld  
(lsb:mysql):    Started mysql-node2



....

...

Jan 22 10:25:46 mysql-node1 rchal: CPU frequency scaling is not supported by 
your processor.

Jan 22 10:25:46 mysql-node1 rchal: boot with 'CPUFREQ=no' in to avoid this 
warning.

Jan 22 10:25:47 mysql-node1 rchal: Cannot load cpufreq governors - No cpufreq 
driver available

Jan 22 10:25:48 mysql-node1 ifup:     lo

Jan 22 10:25:48 mysql-node1 ifup:     lo

Jan 22 10:25:48 mysql-node1 ifup: IP address: 127.0.0.1/8

Jan 22 10:25:48 mysql-node1 ifup:

Jan 22 10:25:48 mysql-node1 ifup:

Jan 22 10:25:48 mysql-node1 ifup: IP address: 127.0.0.2/8

Jan 22 10:25:48 mysql-node1 ifup:

Jan 22 10:25:48 mysql-node1 ifup:     eth0      device: Intel Corporation 
82545EM Gigabit Ethernet Controller (Copper) (rev 01)

Jan 22 10:25:48 mysql-node1 ifup:     eth0

Jan 22 10:25:48 mysql-node1 ifup: IP address: 192.168.96.11/24


Node2 debug log:


Jan 22 10:25:13 mysql-node2 attrd_updater: [21677]: info: attrd_lazy_update: 
Connecting to cluster... 5 retries remaining

Jan 22 10:25:13 mysql-node2 attrd_updater: [21677]: debug: 
init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/attrd

Jan 22 10:25:13 mysql-node2 attrd_updater: [21677]: debug: attrd_update: Sent 
update: pingd=1 for localhost

Jan 22 10:25:13 mysql-node2 attrd: [27805]: debug: attrd_local_callback: update 
message from attrd_updater: pingd=1

Jan 22 10:25:13 mysql-node2 attrd: [27805]: debug: attrd_local_callback: 
Supplied: 1, Current: 1, Stored: 1

Jan 22 10:25:17 mysql-node2 kernel: [104450.241909] block drbd0: PingAck did 
not arrive in time.

Jan 22 10:25:17 mysql-node2 kernel: [104450.241919] block drbd0: peer( 
Secondary -> Unknown ) conn( Connected -> NetworkFailure ) pdsk( UpToDate -> 
DUnknown )

Jan 22 10:25:17 mysql-node2 kernel: [104450.241940] block drbd0: asender 
terminated

Jan 22 10:25:17 mysql-node2 kernel: [104450.241943] block drbd0: Terminating 
asender thread

Jan 22 10:25:17 mysql-node2 kernel: [104450.241989] block drbd0: short read 
expecting header on sock: r=-512

Jan 22 10:25:17 mysql-node2 kernel: [104450.242055] block drbd0: Creating new 
current UUID

Jan 22 10:25:17 mysql-node2 kernel: [104450.242316] block drbd0: Connection 
closed

Jan 22 10:25:17 mysql-node2 kernel: [104450.242322] block drbd0: conn( 
NetworkFailure -> Unconnected )

Jan 22 10:25:17 mysql-node2 kernel: [104450.242326] block drbd0: receiver 
terminated

Jan 22 10:25:17 mysql-node2 kernel: [104450.242328] block drbd0: Restarting 
receiver thread

Jan 22 10:25:17 mysql-node2 kernel: [104450.242330] block drbd0: receiver 
(re)started

Jan 22 10:25:17 mysql-node2 kernel: [104450.242334] block drbd0: conn( 
Unconnected -> WFConnection )

Jan 22 10:25:18 mysql-node2 corosync[27796]:   [TOTEM ] The token was lost in 
the OPERATIONAL state.

Jan 22 10:25:18 mysql-node2 corosync[27796]:   [TOTEM ] A processor failed, 
forming new configuration.

Jan 22 10:25:18 mysql-node2 corosync[27796]:   [TOTEM ] Receive multicast 
socket recv buffer size (262142 bytes).

Jan 22 10:25:18 mysql-node2 corosync[27796]:   [TOTEM ] Transmit multicast 
socket send buffer size (262142 bytes).

Jan 22 10:25:18 mysql-node2 corosync[27796]:   [TOTEM ] entering GATHER state 
from 2.

Jan 22 10:25:23 mysql-node2 lrmd: [27804]: debug: rsc:notification:1:14: monitor

Jan 22 10:25:23 mysql-node2 lrmd: [21678]: debug: perform_ra_op: resetting 
scheduler class to SCHED_OTHER

Jan 22 10:25:23 mysql-node2 lrmd: [27804]: debug: RA output: 
(notification:1:monitor:stdout) running

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] entering GATHER state 
from 0.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Creating commit token 
because I am the rep.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Saving state aru 1d2 
high seq received 1d2

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Storing new sequence id 
for ring e0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] entering COMMIT state.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] got commit token

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] entering RECOVERY state.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] TRANS [0] member 
10.0.0.102:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] position [0] member 
10.0.0.102:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] previous ring seq 220 
rep 10.0.0.101

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] aru 1d2 high delivered 
1d2 received flag 1

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Did not need to 
originate any messages in recovery.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] got commit token

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Sending initial ORF 
token

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] token retrans flag is 0 
my set retrans flag0 retrans queue empty 1 count 0, aru 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] install seq 0 aru 0 
high seq received 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] token retrans flag is 0 
my set retrans flag0 retrans queue empty 1 count 1, aru 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] install seq 0 aru 0 
high seq received 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] token retrans flag is 0 
my set retrans flag0 retrans queue empty 1 count 2, aru 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] install seq 0 aru 0 
high seq received 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] token retrans flag is 0 
my set retrans flag0 retrans queue empty 1 count 3, aru 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] install seq 0 aru 0 
high seq received 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] retrans flag count 4 
token aru 0 install seq 0 aru 0 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Resetting old ring state

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] recovery to regular 1-0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Delivering to app 1d3 
to 1d2

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] CLM CONFIGURATION CHANGE

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] New Configuration:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ]         r(0) 
ip(10.0.0.102)

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] Members Left:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ]         r(0) 
ip(10.0.0.101)

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] Members Joined:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [EVT   ] Evt conf change 1

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [EVT   ] m 1, j 0 l 1

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [LCK   ] [DEBUG]: lck_confchg_fn

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [MSG   ] [DEBUG]: msg_confchg_fn

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] notice: 
pcmk_peer_update: Transitional membership event on ring 224: memb=1, new=0, 
lost=1

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] info: pcmk_peer_update: 
memb: mysql-node2 1711276042

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] info: pcmk_peer_update: 
lost: mysql-node1 1694498826

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] CLM CONFIGURATION CHANGE

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] New Configuration:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ]         r(0) 
ip(10.0.0.102)

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] Members Left:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [CLM   ] Members Joined:

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [EVT   ] Evt conf change 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [EVT   ] m 1, j 0 l 0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [LCK   ] [DEBUG]: lck_confchg_fn

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [MSG   ] [DEBUG]: msg_confchg_fn

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] notice: 
pcmk_peer_update: Stable membership event on ring 224: memb=1, new=0, lost=0

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] info: pcmk_peer_update: 
MEMB: mysql-node2 1711276042

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] info: 
ais_mark_unseen_peer_dead: Node mysql-node1 was not seen in the previous 
transition

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] info: update_member: 
Node 1694498826/mysql-node1 is now: lost

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] debug: 
pcmk_peer_update: 1 nodes changed

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] info: 
send_member_notification: Sending membership update 224 to 2 children

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] debug: send_cluster_id: 
Local update: id=1711276042, born=220, seq=224

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [SYNC  ] This node is within the 
primary component and will provide service.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] entering OPERATIONAL 
state.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] A processor joined or 
left the membership and a new membership was formed.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] mcasted message added 
to pending queue

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] mcasted message added 
to pending queue

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Delivering 0 to 2

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [TOTEM ] Delivering MCAST 
message with seq 1 to pending delivery queue

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [pcmk  ] debug: 
pcmk_cluster_id_callback: Node update: mysql-node2 (1.1.2)

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [SYNC  ] confchg entries 1

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [SYNC  ] Barrier Start Received 
From 1711276042

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [SYNC  ] Barrier completion 
status for nodeid 1711276042 = 1.

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [SYNC  ] Synchronization barrier 
completed

Jan 22 10:25:26 mysql-node2 corosync[27796]:   [SYNC  ] Synchronization actions 
starting for (openais cluster membership service B.01.01)

...

...


Viele Grüße aus Neustadt,

Thomas Fachtan
Database Administrator

Fon: +49 9602 9442-301
Fax: +49 9602 9442-309
E-Mail: [email protected]
.................................................
motiondrive AG
Im Wiesengrund 1
92660 Neustadt
http://www.motiondrive.ag

Vorstandsvorsitzender: Werner A. Krachtus
Aufsichtsratsvorsitzender: Dipl.-Kfm. Ralf Barth
Handelsregister Weiden: HRB 2817

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to