Hello,

I have created a second ring in corosync.cfg on my two nodes.

# This specifies the mode of redundant ring, which may be none, active, or 
passive.
        rrp_mode: active

        interface {
                # The following values need to be set based on your environment
                ringnumber: 0
                bindnetaddr: 192.168.138.0
                mcastaddr: 226.94.1.1
                mcastport: 5405
        }
        interface {
                # The following values need to be set based on your environment
                ringnumber: 1
                bindnetaddr: 220.0.0.0
                mcastaddr: 226.94.1.2
                mcastport: 5415
        }

Before activating the config by restarting the corosync service, both nodes 
worked in a active passive cluster.
But now the second node (lxds07) is marked in crm_mon as offline.

In the corosync.log appear such entries:

Jan 03 12:43:59 lxds07 crmd: [1364]: info: do_state_transition: State 
transition S_FINALIZE_JOIN -> S_INTEGRATION [ input=I_ELECTION_DC 
cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
Jan 03 12:43:59 lxds07 crmd: [1364]: info: do_dc_join_offer_all: join-13: 
Waiting on 2 outstanding join acks
Jan 03 12:46:59 lxds07 crmd: [1364]: ERROR: crm_timer_popped: Integration Timer 
(I_INTEGRATED) just popped!
Jan 03 12:46:59 lxds07 crmd: [1364]: info: crm_timer_popped: Welcomed: 2, 
Integrated: 0
Jan 03 12:46:59 lxds07 crmd: [1364]: info: do_state_transition: State 
transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED 
cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 03 12:46:59 lxds07 crmd: [1364]: WARN: do_state_transition: Progressed to 
state S_FINALIZE_JOIN after C_TIMER_POPPED
Jan 03 12:46:59 lxds07 crmd: [1364]: WARN: do_state_transition: 2 cluster nodes 
failed to respond to the join offer.
Jan 03 12:46:59 lxds07 crmd: [1364]: info: ghash_print_node:   Welcome reply 
not received from: lxds05 13
Jan 03 12:46:59 lxds07 crmd: [1364]: info: ghash_print_node:   Welcome reply 
not received from: lxds07 13
Jan 03 12:46:59 lxds07 crmd: [1364]: WARN: do_log: FSA: Input I_ELECTION_DC 
from do_dc_join_finalize() received in state S_FINALIZE_JOIN
Jan 03 12:46:59 lxds07 crmd: [1364]: info: do_state_transition: State 
transition S_FINALIZE_JOIN -> S_INTEGRATION [ input=I_ELECTION_DC 
cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
Jan 03 12:46:59 lxds07 crmd: [1364]: info: do_dc_join_offer_all: join-14: 
Waiting on 2 outstanding join acks
Jan 03 12:49:59 lxds07 crmd: [1364]: ERROR: crm_timer_popped: Integration Timer 
(I_INTEGRATED) just popped!
Jan 03 12:49:59 lxds07 crmd: [1364]: info: crm_timer_popped: Welcomed: 2, 
Integrated: 0
Jan 03 12:49:59 lxds07 crmd: [1364]: info: do_state_transition: State 
transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED 
cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 03 12:49:59 lxds07 crmd: [1364]: WARN: do_state_transition: Progressed to 
state S_FINALIZE_JOIN after C_TIMER_POPPED
Jan 03 12:49:59 lxds07 crmd: [1364]: WARN: do_state_transition: 2 cluster nodes 
failed to respond to the join offer.
Jan 03 12:49:59 lxds07 crmd: [1364]: info: ghash_print_node:   Welcome reply 
not received from: lxds05 14
Jan 03 12:49:59 lxds07 crmd: [1364]: info: ghash_print_node:   Welcome reply 
not received from: lxds07 14
Jan 03 12:49:59 lxds07 crmd: [1364]: WARN: do_log: FSA: Input I_ELECTION_DC 
from do_dc_join_finalize() received in state S_FINALIZE_JOIN
Jan 03 12:49:59 lxds07 crmd: [1364]: info: do_state_transition: State 
transition S_FINALIZE_JOIN -> S_INTEGRATION [ input=I_ELECTION_DC 
cause=C_FSA_INTERNAL origin=do_dc_join_finalize ]
Jan 03 12:49:59 lxds07 crmd: [1364]: info: do_dc_join_offer_all: join-15: 
Waiting on 2 outstanding join acks

root@lxds07:/home/adminsen# corosync-cfgtool -s
Printing ring status.
Local node ID 1401596096
RING ID 0
        id      = 192.168.138.83
        status  = ring 0 active with no faults
RING ID 1
        id      = 220.0.0.3
        status  = ring 1 active with no faults


root@lxds05:/home/adminsen# corosync-cfgtool -s
Printing ring status.
Local node ID 1821026496
RING ID 0
        id      = 192.168.138.108
        status  = ring 0 active with no faults
RING ID 1
        id      = 220.0.0.2
        status  = ring 1 active with no faults


Can you please give me a hint, what is wrong with my configuration?


Regards and best wishes for the new year!

Stefan

<<attachment: winmail.dat>>

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to