Gather from state 3 back to back is an indicator that iptables are not
properly configured on the node. I know you said iptables are turned
off, but if iptables are off, the node would at least form a singleton ring.
Could you send your config file?
Regards
-steve
On 01/26/2011 09:01 AM, Dan Frincu wrote:
> Update: increased verbosity to debug and I get the following
>
> Jan 26 16:36:45 corosync [TOTEM ] The consensus timeout expired.
> Jan 26 16:36:45 corosync [TOTEM ] entering GATHER state from 3.
> Jan 26 16:36:53 corosync [TOTEM ] The consensus timeout expired.
> Jan 26 16:36:53 corosync [TOTEM ] entering GATHER state from 3.
> Jan 26 16:37:00 corosync [TOTEM ] The consensus timeout expired.
> Jan 26 16:37:00 corosync [TOTEM ] entering GATHER state from 3.
> Jan 26 16:37:08 corosync [TOTEM ] The consensus timeout expired.
> Jan 26 16:37:08 corosync [TOTEM ] entering GATHER state from 3.
> Jan 26 16:37:12 cluster1 crmd: [16266]: ERROR: crm_timer_popped:
> Integration Timer (I_INTEGRATED) just popped!
> Jan 26 16:37:12 cluster1 crmd: [16266]: info: crm_timer_popped:
> Welcomed: 1, Integrated: 0
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: s_crmd_fsa: Processing
> I_INTEGRATED: [ state=S_INTEGRATION cause=C_TIMER_POPPED
> origin=crm_timer_popped ]
> Jan 26 16:37:12 cluster1 crmd: [16266]: info: do_state_transition: State
> transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED
> cause=C_TIMER_POPPED origin=crm_timer_popped ]
> Jan 26 16:37:12 cluster1 crmd: [16266]: WARN: do_state_transition:
> Progressed to state S_FINALIZE_JOIN after C_TIMER_POPPED
> Jan 26 16:37:12 cluster1 crmd: [16266]: WARN: do_state_transition: 1
> cluster nodes failed to respond to the join offer.
> Jan 26 16:37:12 cluster1 crmd: [16266]: info: ghash_print_node:
> Welcome reply not received from: cluster1 42
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_DC_TIMER_STOP
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_INTEGRATE_TIMER_STOP
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_FINALIZE_TIMER_START
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: crm_timer_start: Started
> Finalization Timer (I_ELECTION:1800000ms), src=102
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_DC_JOIN_FINALIZE
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_dc_join_finalize:
> Finializing join-42 for 0 clients
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: s_crmd_fsa: Processing
> I_ELECTION_DC: [ state=S_FINALIZE_JOIN cause=C_FSA_INTERNAL
> origin=do_dc_join_finalize ]
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_WARN
> Jan 26 16:37:12 cluster1 crmd: [16266]: WARN: do_log: FSA: Input
> I_ELECTION_DC from do_dc_join_finalize() received in state S_FINALIZE_JOIN
> Jan 26 16:37:12 cluster1 crmd: [16266]: info: do_state_transition: State
> transition S_FINALIZE_JOIN -> S_INTEGRATION [ input=I_ELECTION_DC
> cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_DC_TIMER_STOP
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_INTEGRATE_TIMER_START
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: crm_timer_start: Started
> Integration Timer (I_INTEGRATED:180000ms), src=103
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_FINALIZE_TIMER_STOP
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_ELECTION_VOTE
> Jan 26 16:37:12 corosync [TOTEM ] mcasted message added to pending queue
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_election_vote: Started
> election 44
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: do_fsa_action:
> actions:trace: // A_DC_JOIN_OFFER_ALL
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: initialize_join: join-43:
> Initializing join data (flag=true)
> Jan 26 16:37:12 cluster1 crmd: [16266]: debug: join_make_offer: join-43:
> Sending offer to cluster1
> Jan 26 16:37:12 cluster1 crmd: [16266]: info: do_dc_join_offer_all:
> join-43: Waiting on 1 outstanding join acks
> Jan 26 16:37:15 corosync [TOTEM ] The consensus timeout expired.
> Jan 26 16:37:16 corosync [TOTEM ] entering GATHER state from 3.
> Jan 26 16:37:23 corosync [TOTEM ] The consensus timeout expired.
> Jan 26 16:37:23 corosync [TOTEM ] entering GATHER state from 3.
> Jan 26 16:37:31 corosync [TOTEM ] The consensus timeout expired.
>
> Running corosync-blackbox gives me:
>
> # corosync-blackbox
> Starting replay: head [67420] tail [0]
> rec=[1] Log Message=Corosync Cluster Engine ('1.3.0'): started and ready
> to provide service.
> rec=[2] Log Message=Corosync built-in features: nss rdma
> rec=[3] Log Message=Successfully read main configuration file
> '/etc/corosync/corosync.conf'.
> rec=[4] Log Message=Token Timeout (5000 ms) retransmit timeout (490 ms)
> rec=[5] Log Message=token hold (382 ms) retransmits before loss (10 retrans)
> rec=[6] Log Message=join (1000 ms) send_join (45 ms) consensus (2500 ms)
> merge (200 ms)
> rec=[7] Log Message=downcheck (1000 ms) fail to recv const (50 msgs)
> rec=[8] Log Message=seqno unchanged const (30 rotations) Maximum network
> MTU 1402
> rec=[9] Log Message=window size per rotation (50 messages) maximum
> messages per rotation (25 messages)
> rec=[10] Log Message=send threads (0 threads)
> rec=[11] Log Message=RRP token expired timeout (490 ms)
> rec=[12] Log Message=RRP token problem counter (2000 ms)
> rec=[13] Log Message=RRP threshold (10 problem count)
> rec=[14] Log Message=RRP mode set to active.
> rec=[15] Log Message=heartbeat_failures_allowed (0)
> rec=[16] Log Message=max_network_delay (50 ms)
> rec=[17] Log Message=HeartBeat is Disabled. To enable set
> heartbeat_failures_allowed > 0
> rec=[18] Log Message=Initializing transport (UDP/IP Unicast).
> rec=[19] Log Message=Initializing transmit/receive security: libtomcrypt
> SOBER128/SHA1HMAC (mode 0).
> rec=[20] Log Message=Initializing transport (UDP/IP Unicast).
> rec=[21] Log Message=Initializing transmit/receive security: libtomcrypt
> SOBER128/SHA1HMAC (mode 0).
> rec=[22] Log Message=you are using ipc api v2
> rec=[23] Log Message=The network interface [10.0.2.11] is now up.
> rec=[24] Log Message=Created or loaded sequence id 0.10.0.2.11 for this
> ring.
> rec=[25] Log Message=debug: pcmk_user_lookup: Cluster user root has
> uid=0 gid=0
> rec=[26] Log Message=info: process_ais_conf: Reading configure
> rec=[27] Log Message=info: config_find_init: Local handle:
> 4552499517957603332 for logging
> rec=[28] Log Message=info: config_find_next: Processing additional
> logging options...
> rec=[29] Log Message=info: get_config_opt: Found 'on' for option: debug
> rec=[30] Log Message=info: get_config_opt: Found 'yes' for option:
> to_logfile
> rec=[31] Log Message=info: get_config_opt: Found
> '/var/log/cluster/corosync.log' for option: logfile
> rec=[32] Log Message=info: get_config_opt: Found 'no' for option: to_syslog
> rec=[33] Log Message=info: process_ais_conf: User configured file based
> logging and explicitly disabled syslog.
> rec=[34] Log Message=info: config_find_init: Local handle:
> 8972265949260414981 for service
> rec=[35] Log Message=info: config_find_next: Processing additional
> service options...
> rec=[36] Log Message=info: get_config_opt: Defaulting to 'pcmk' for
> option: clustername
> rec=[37] Log Message=info: get_config_opt: Found 'no' for option: use_logd
> rec=[38] Log Message=info: get_config_opt: Found 'yes' for option: use_mgmtd
> rec=[39] Log Message=info: pcmk_startup: CRM: Initialized
> rec=[40] Log Message=Logging: Initialized pcmk_startup
> rec=[41] Log Message=info: pcmk_startup: Maximum core file size is:
> 18446744073709551615
> rec=[42] Log Message=debug: pcmk_user_lookup: Cluster user hacluster has
> uid=101 gid=102
> rec=[43] Log Message=info: pcmk_startup: Service: 9
> rec=[44] Log Message=info: pcmk_startup: Local hostname: cluster1
> rec=[45] Log Message=info: pcmk_update_nodeid: Local node id: 184680458
> rec=[46] Log Message=info: update_member: Creating entry for node
> 184680458 born on 0
> rec=[47] Log Message=info: update_member: 0x5fe73e0 Node 184680458 now
> known as cluster1 (was: (null))
> rec=[48] Log Message=info: update_member: Node cluster1 now has 1 quorum
> votes (was 0)
> rec=[49] Log Message=info: update_member: Node 184680458/cluster1 is
> now: member
> rec=[50] Log Message=info: spawn_child: Forked child 16261 for process
> stonithd
> rec=[51] Log Message=debug: pcmk_user_lookup: Cluster user hacluster has
> uid=101 gid=102
> rec=[52] Log Message=info: spawn_child: Forked child 16262 for process cib
> rec=[53] Log Message=info: spawn_child: Forked child 16263 for process lrmd
> rec=[54] Log Message=debug: pcmk_user_lookup: Cluster user hacluster has
> uid=101 gid=102
> rec=[55] Log Message=info: spawn_child: Forked child 16264 for process attrd
> rec=[56] Log Message=debug: pcmk_user_lookup: Cluster user hacluster has
> uid=101 gid=102
> rec=[57] Log Message=info: spawn_child: Forked child 16265 for process
> pengine
> rec=[58] Log Message=debug: pcmk_user_lookup: Cluster user hacluster has
> uid=101 gid=102
> rec=[59] Log Message=info: spawn_child: Forked child 16266 for process crmd
> *rec=[60] Log Message=spawn_child: FATAL: Cannot exec
> /usr/lib64/heartbeat/mgmtd: (2) No such file or directory*
> *** buffer overflow detected ***: corosync-fplay terminated
> ======= Backtrace: =========
> /lib64/libc.so.6(__chk_fail+0x2f)[0x37a72e6c2f]
> corosync-fplay[0x400c0b]
> /lib64/libc.so.6(__libc_start_main+0xf4)[0x37a721d974]
> corosync-fplay[0x4008d9]
> ======= Memory map: ========
> 00400000-00402000 r-xp 00000000 08:05 1041788
> /usr/sbin/corosync-fplay
> 00602000-00603000 rw-p 00002000 08:05 1041788
> /usr/sbin/corosync-fplay
> 00603000-0060d000 rw-p 00603000 00:00 0
> 0af45000-0af66000 rw-p 0af45000 00:00 0
> [heap]
> 3190400000-319040d000 r-xp 00000000 08:02 911558
> /lib64/libgcc_s-4.1.2-20080825.so.1
> 319040d000-319060d000 ---p 0000d000 08:02 911558
> /lib64/libgcc_s-4.1.2-20080825.so.1
> 319060d000-319060e000 rw-p 0000d000 08:02 911558
> /lib64/libgcc_s-4.1.2-20080825.so.1
> 37a6e00000-37a6e1c000 r-xp 00000000 08:02 911525
> /lib64/ld-2.5.so <http://ld-2.5.so>
> 37a701b000-37a701c000 r--p 0001b000 08:02 911525
> /lib64/ld-2.5.so <http://ld-2.5.so>
> 37a701c000-37a701d000 rw-p 0001c000 08:02 911525
> /lib64/ld-2.5.so <http://ld-2.5.so>
> 37a7200000-37a734c000 r-xp 00000000 08:02 911526
> /lib64/libc-2.5.so <http://libc-2.5.so>
> 37a734c000-37a754c000 ---p 0014c000 08:02 911526
> /lib64/libc-2.5.so <http://libc-2.5.so>
> 37a754c000-37a7550000 r--p 0014c000 08:02 911526
> /lib64/libc-2.5.so <http://libc-2.5.so>
> 37a7550000-37a7551000 rw-p 00150000 08:02 911526
> /lib64/libc-2.5.so <http://libc-2.5.so>
> 37a7551000-37a7556000 rw-p 37a7551000 00:00 0
> 37a7600000-37a7602000 r-xp 00000000 08:02 911527
> /lib64/libdl-2.5.so <http://libdl-2.5.so>
> 37a7602000-37a7802000 ---p 00002000 08:02 911527
> /lib64/libdl-2.5.so <http://libdl-2.5.so>
> 37a7802000-37a7803000 r--p 00002000 08:02 911527
> /lib64/libdl-2.5.so <http://libdl-2.5.so>
> 37a7803000-37a7804000 rw-p 00003000 08:02 911527
> /lib64/libdl-2.5.so <http://libdl-2.5.so>
> 37a7e00000-37a7e16000 r-xp 00000000 08:02 911531
> /lib64/libpthread-2.5.so <http://libpthread-2.5.so>
> 37a7e16000-37a8015000 ---p 00016000 08:02 911531
> /lib64/libpthread-2.5.so <http://libpthread-2.5.so>
> 37a8015000-37a8016000 r--p 00015000 08:02 911531
> /lib64/libpthread-2.5.so <http://libpthread-2.5.so>
> 37a8016000-37a8017000 rw-p 00016000 08:02 911531
> /lib64/libpthread-2.5.so <http://libpthread-2.5.so>
> 37a8017000-37a801b000 rw-p 37a8017000 00:00 0
> 37a8e00000-37a8e07000 r-xp 00000000 08:02 911532
> /lib64/librt-2.5.so <http://librt-2.5.so>
> 37a8e07000-37a9007000 ---p 00007000 08:02 911532
> /lib64/librt-2.5.so <http://librt-2.5.so>
> 37a9007000-37a9008000 r--p 00007000 08:02 911532
> /lib64/librt-2.5.so <http://librt-2.5.so>
> 37a9008000-37a9009000 rw-p 00008000 08:02 911532
> /lib64/librt-2.5.so <http://librt-2.5.so>
> 2b3bae55d000-2b3bae55e000 rw-p 2b3bae55d000 00:00 0
> 2b3bae56a000-2b3bae943000 rw-p 2b3bae56a000 00:00 0
> 7ffffc537000-7ffffc54c000 rw-p 7ffffffea000 00:00 0
> [stack]
> ffffffffff600000-ffffffffffe00000 ---p 00000000 00:00 0
> [vdso]
> /usr/bin/corosync-blackbox: line 34: 16676 Aborted
> corosync-fplay
>
> I see the error message with mgmtd however I've performed the same test
> on a pair of XEN VM's with the exact same packages (clean install, no
> upgrade from openais-0.80 like the real hardware) and mgmtd doesn't
> exist either, but it says
>
> rec=[55] Log Message=info: spawn_child: Forked child 4459 for process mgmtd
>
> # ll /usr/lib64/heartbeat/mgmtd
> ls: /usr/lib64/heartbeat/mgmtd: No such file or directory
>
> # rpm -ql pacemaker-1.0.10-1.4 | grep mgm
> /usr/lib64/python2.4/site-packages/crm/idmgmt.py
> /usr/lib64/python2.4/site-packages/crm/idmgmt.pyc
> /usr/lib64/python2.4/site-packages/crm/idmgmt.pyo
>
> Anyone?
>
> Regards,
> Dan
>
> On Wed, Jan 26, 2011 at 1:35 PM, Dan Frincu <[email protected]
> <mailto:[email protected]>> wrote:
>
> Hi,
>
> I've got a pair of servers running on RHEL5 x86_64 with openais-0.80
> (older install) which I want to upgrade to corosync-1.3.0 +
> pacemaker-1.0.10. Downtime is not an issue and corosync 1.3.0 is
> needed for UDPU, so I built it from the corosync.org
> <http://corosync.org/> website.
>
> With pacemaker, we won't be using the heartbeat stack, so I built
> the pacemaker package from the clusterlabs.org
> <http://clusterlabs.org/> src.rpm without heartbeat support. To be
> more precise I used
>
> rpmbuild --without heartbeat --with ais --with snmp --with esmtp -ba
> pacemaker-epel.spec
>
> Now I've tested the rpm list below on a pair of XEN VM's, it works
> just fine.
>
> cluster-glue-1.0.6-1.6.el5.x86_64.rpm
> cluster-glue-libs-1.0.6-1.6.el5.x86_64.rpm
> corosync-1.3.0-1.x86_64.rpm
> corosynclib-1.3.0-1.x86_64.rpm
> libesmtp-1.0.4-5.el5.x86_64.rpm
> libibverbs-1.1.2-1.el5.x86_64.rpm
> librdmacm-1.0.8-1.el5.x86_64.rpm
> libtool-ltdl-1.5.22-6.1.x86_64.rpm
> openais-1.1.4-2.x86_64.rpm
> openaislib-1.1.4-2.x86_64.rpm
> openhpi-2.10.2-1.el5.x86_64.rpm
> openib-1.3.2-0.20080728.0355.3.el5.noarch.rpm
> pacemaker-1.0.10-1.4.x86_64.rpm
> pacemaker-libs-1.0.10-1.4.x86_64.rpm
> perl-TimeDate-1.16-5.el5.noarch.rpm
> resource-agents-1.0.3-2.6.el5.x86_64.rpm
>
> However when performing the upgrade on the servers running
> openais-0.80, first I removed the heartbeat, heartbeat-libs and
> PyXML rpms (conflicting dependencies issue) then rpm -Uvh on the rpm
> list above. Installation went fine, removed existing cib.xml and
> signatures, fresh start. Then I configured corosync, then started it
> on both servers, and nothing. At first I got an error related to
> pacemaker mgmt, which was an old package installed with the old
> rpms. Removed it, tried again. Nothing. Removed all cluster related
> rpms old and new + deps, except for DRBD, then installed the list
> above, then again, nothing. What nothing means:
> - corosync starts, never elects DC, never sees the other node or
> itself for that matter.
> - can stop corosync via the init script, it goes into an endless
> phase where it just prints dots to the screen, have to kill the
> process to make it stop.
>
> Troubleshooting done so far:
> - tested network sockets (nc from side to side), firewall rules
> (iptables down), communication is ok
> - searched for the original RPM's list, removed all remaining RPMs,
> ran ldconfig, removed new RPM's, installed new RPM's
>
> My guess is that there are some leftovers from the old openais-0.80
> installation, which mess with the current installation, seeing as
> how the same set of RPMs on a pair of XEN VM's with the same OS work
> fine, however I cannot put my finger on the culprit for the real
> servers' issue.
>
> Logs: http://pastebin.com/i0maZM4p
>
> Removed everything after removing the RPM's, just to be extra
> paranoid about leftovers (rpm -qpl *.rpm >> file && for i in `cat
> file `; do [[ -e "$i" ]] && echo "$i" >> newfile ; done && for i in
> `cat newfile` ; do rm -rf $i ; done)
>
> Installed RPMs (without openais)
>
> Same output
>
> http://pastebin.com/3iPHSXua
>
> It seems to go into some sort of loop.
>
> Jan 26 12:13:41 cluster1 crmd: [15612]: ERROR: crm_timer_popped:
> Integration Timer (I_INTEGRATED) just popped!
> Jan 26 12:13:41 cluster1 crmd: [15612]: info: crm_timer_popped:
> Welcomed: 1, Integrated: 0
> Jan 26 12:13:41 cluster1 crmd: [15612]: info: do_state_transition:
> State transition S_INTEGRATION -> S_FINALIZE_JOIN [
> input=I_INTEGRATED cause=C_TIMER_POPPED origin=crm_timer_popped ]
> Jan 26 12:13:41 cluster1 crmd: [15612]: WARN: do_state_transition:
> Progressed to state S_FINALIZE_JOIN after C_TIMER_POPPED
> Jan 26 12:13:41 cluster1 crmd: [15612]: WARN: do_state_transition: 1
> cluster nodes failed to respond to the join offer.
> Jan 26 12:13:41 cluster1 crmd: [15612]: info: ghash_print_node:
> Welcome reply not received from: cluster1 7
> Jan 26 12:13:41 cluster1 crmd: [15612]: WARN: do_log: FSA: Input
> I_ELECTION_DC from do_dc_join_finalize() received in state
> S_FINALIZE_JOIN
> Jan 26 12:13:41 cluster1 crmd: [15612]: info: do_state_transition:
> State transition S_FINALIZE_JOIN -> S_INTEGRATION [
> input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
> Jan 26 12:13:41 cluster1 crmd: [15612]: info: do_dc_join_offer_all:
> join-8: Waiting on 1 outstanding join acks
> Jan 26 12:16:41 cluster1 crmd: [15612]: ERROR: crm_timer_popped:
> Integration Timer (I_INTEGRATED) just popped!
> Jan 26 12:16:41 cluster1 crmd: [15612]: info: crm_timer_popped:
> Welcomed: 1, Integrated: 0
> Jan 26 12:16:41 cluster1 crmd: [15612]: info: do_state_transition:
> State transition S_INTEGRATION -> S_FINALIZE_JOIN [
> input=I_INTEGRATED cause=C_TIMER_POPPED origin=crm_timer_popped ]
> Jan 26 12:16:41 cluster1 crmd: [15612]: WARN: do_state_transition:
> Progressed to state S_FINALIZE_JOIN after C_TIMER_POPPED
> Jan 26 12:16:41 cluster1 crmd: [15612]: WARN: do_state_transition: 1
> cluster nodes failed to respond to the join offer.
> Jan 26 12:16:41 cluster1 crmd: [15612]: info: ghash_print_node:
> Welcome reply not received from: cluster1 8
> Jan 26 12:16:41 cluster1 crmd: [15612]: WARN: do_log: FSA: Input
> I_ELECTION_DC from do_dc_join_finalize() received in state
> S_FINALIZE_JOIN
> Jan 26 12:16:41 cluster1 crmd: [15612]: info: do_state_transition:
> State transition S_FINALIZE_JOIN -> S_INTEGRATION [
> input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
> Jan 26 12:16:41 cluster1 crmd: [15612]: info: do_dc_join_offer_all:
> join-9: Waiting on 1 outstanding join acks
> Jan 26 12:19:41 cluster1 crmd: [15612]: ERROR: crm_timer_popped:
> Integration Timer (I_INTEGRATED) just popped!
> Jan 26 12:19:41 cluster1 crmd: [15612]: info: crm_timer_popped:
> Welcomed: 1, Integrated: 0
> Jan 26 12:19:41 cluster1 crmd: [15612]: info: do_state_transition:
> State transition S_INTEGRATION -> S_FINALIZE_JOIN [
> input=I_INTEGRATED cause=C_TIMER_POPPED origin=crm_timer_popped ]
> Jan 26 12:19:41 cluster1 crmd: [15612]: WARN: do_state_transition:
> Progressed to state S_FINALIZE_JOIN after C_TIMER_POPPED
> Jan 26 12:19:41 cluster1 crmd: [15612]: WARN: do_state_transition: 1
> cluster nodes failed to respond to the join offer.
> Jan 26 12:19:41 cluster1 crmd: [15612]: info: ghash_print_node:
> Welcome reply not received from: cluster1 9
> Jan 26 12:19:41 cluster1 crmd: [15612]: WARN: do_log: FSA: Input
> I_ELECTION_DC from do_dc_join_finalize() received in state
> S_FINALIZE_JOIN
> Jan 26 12:19:41 cluster1 crmd: [15612]: info: do_state_transition:
> State transition S_FINALIZE_JOIN -> S_INTEGRATION [
> input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
> Jan 26 12:19:41 cluster1 crmd: [15612]: info: do_dc_join_offer_all:
> join-10: Waiting on 1 outstanding join acks
> Jan 26 12:20:11 cluster1 cib: [15608]: info: cib_stats: Processed 1
> operations (0.00us average, 0% utilization) in the last 10min
>
> Any suggestions?
>
> TIA.
>
> Regards,
> Dan
>
> --
> Dan Frîncu
> CCNA, RHCE
>
>
>
>
> --
> Dan Frîncu
> CCNA, RHCE
>
>
>
> _______________________________________________
> Openais mailing list
> [email protected]
> https://lists.linux-foundation.org/mailman/listinfo/openais
_______________________________________________
Openais mailing list
[email protected]
https://lists.linux-foundation.org/mailman/listinfo/openais