Hi ,

The sync request from SC2 clearly reaches SC1.
Is any sync started at SC1 ?
I cant see because the syslog snippet from SC1 is minimal, truncated 
right after the request arrives.

/Anders Bjornerstedt

Yao Cheng LIANG wrote:
> Dear all,
>
> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message 
> appears in /var/log/message on sc-1:
> ------------------------------------------------------
> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id  328207 : EE_ID 
> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1
> Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on STANDBY 
> Controller at 5020f
> Jul 12 22:35:26 localhost osafimmd[11706]: IMMND on controller (not currently 
> coord) requests sync
> Jul 12 22:35:26 localhost osafimmd[11706]: Node 5020f request sync 
> sync-pid:8930 epoch:0
> ------------------------------------------------------------------------------------------------------------------------------
>
> while on sc-2, below message appears in /var/log/message:
> ------------------------------------------------------------------------------------------------------------------------------
> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services
> Jul 12 22:35:26 WR20-64_32 osafdtmd[8860]: Started
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup 
> - OK
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK
> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started
>
> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup 
> - OK
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK
> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has 
> active state => Standby role
> Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: Started
> Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : 
> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd startup 
> - OK
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK
> Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Started
> Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Initialization Success, role 
> STANDBY
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmd startup 
> - OK
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Started
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Initialization Success
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Director Service is up
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd 
> startup - OK
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: 
> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: 
> IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: 
> IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING
> Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: NODE STATE-> IMM_NODE_ISOLATED
> Jul 12 22:35:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 20 seconds
> Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 40 seconds
> Jul 12 22:36:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 60 seconds
> Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 80 seconds
> Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 1000
> Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 100 seconds
> Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync request, when 
> IMM_NODE_ISOLATED
> Jul 12 22:37:16 WR20-64_32 osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed 
> rc : 22
> Jul 12 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 120 seconds
> Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 140 seconds
> Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 
> detected at standby immd!! f2. Possible failover
> Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will 
> not mbcp to peer IMMD
> Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message count:1856 + 1 != 1856
> Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 160 seconds
> Jul 12 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 180 seconds
> Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 2000
> Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 200 seconds
> Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when 
> IMM_NODE_ISOLATED
> Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add New incoming connection 
> to fd : 22
> Jul 12 22:39:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 220 seconds
> Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 240 seconds
> Jul 12 22:39:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 260 seconds
> Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 280 seconds
> Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 3000
> Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 300 seconds
> Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync request, when 
> IMM_NODE_ISOLATED
> -------------------------------------------------------------------------------------------------------------------------------
>
> But I reverse the order - i.e. to start sc-2 and then sc-2, both controller 
> can be started successfully
>
> Could anyone tell me what’s wrong?
>
> Thanks.
>
> Ted
>
>
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> This message (including any attachments) is for the named
> addressee(s)'s use only. It may contain sensitive, confidential,
> private proprietary or legally privileged information intended for a
> specific individual and purpose, and is protected by law. If you are
> not the intended recipient, please immediately delete it and all copies
> of it from your system, destroy any hard copies of it
> and notify the sender. Any use, disclosure, copying, or distribution of
> this message and/or any attachments is strictly prohibited.
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>
> ------------------------------------------------------------------------------
> _______________________________________________
> Opensaf-users mailing list
> [email protected]
> https://lists.sourceforge.net/lists/listinfo/opensaf-users
>   



------------------------------------------------------------------------------
Want fast and easy access to all the code in your enterprise? Index and
search up to 200,000 lines of code with a free copy of Black Duck&#174;
Code Sight&#153; - the same software that powers the world's largest code
search on Ohloh, the Black Duck Open Hub! Try it now.
http://p.sf.net/sfu/bds
_______________________________________________
Opensaf-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/opensaf-users

Reply via email to