Hi , The sync request from SC2 clearly reaches SC1. Is any sync started at SC1 ? I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives.
/Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear all, > > I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message > appears in /var/log/message on sc-1: > ------------------------------------------------------ > Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : EE_ID > safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on STANDBY > Controller at 5020f > Jul 12 22:35:26 localhost osafimmd[11706]: IMMND on controller (not currently > coord) requests sync > Jul 12 22:35:26 localhost osafimmd[11706]: Node 5020f request sync > sync-pid:8930 epoch:0 > ------------------------------------------------------------------------------------------------------------------------------ > > while on sc-2, below message appears in /var/log/message: > ------------------------------------------------------------------------------------------------------------------------------ > Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services > Jul 12 22:35:26 WR20-64_32 osafdtmd[8860]: Started > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup > - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started > > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup > - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has > active state => Standby role > Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: Started > Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : > safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd startup > - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Started > Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Initialization Success, role > STANDBY > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmd startup > - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Started > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Initialization Success > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Director Service is up > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd > startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: > IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: > IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: > IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: NODE STATE-> IMM_NODE_ISOLATED > Jul 12 22:35:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 20 seconds > Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 40 seconds > Jul 12 22:36:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 60 seconds > Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 80 seconds > Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 > Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 100 seconds > Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync request, when > IMM_NODE_ISOLATED > Jul 12 22:37:16 WR20-64_32 osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed > rc : 22 > Jul 12 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 120 seconds > Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 140 seconds > Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 > detected at standby immd!! f2. Possible failover > Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will > not mbcp to peer IMMD > Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message count:1856 + 1 != 1856 > Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 160 seconds > Jul 12 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 180 seconds > Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 2000 > Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 200 seconds > Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when > IMM_NODE_ISOLATED > Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add New incoming connection > to fd : 22 > Jul 12 22:39:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 220 seconds > Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 240 seconds > Jul 12 22:39:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 260 seconds > Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 280 seconds > Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 > Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 300 seconds > Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync request, when > IMM_NODE_ISOLATED > ------------------------------------------------------------------------------------------------------------------------------- > > But I reverse the order - i.e. to start sc-2 and then sc-2, both controller > can be started successfully > > Could anyone tell me what’s wrong? > > Thanks. > > Ted > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all copies > of it from your system, destroy any hard copies of it > and notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ------------------------------------------------------------------------------ > _______________________________________________ > Opensaf-users mailing list > [email protected] > https://lists.sourceforge.net/lists/listinfo/opensaf-users > ------------------------------------------------------------------------------ Want fast and easy access to all the code in your enterprise? Index and search up to 200,000 lines of code with a free copy of Black Duck® Code Sight™ - the same software that powers the world's largest code search on Ohloh, the Black Duck Open Hub! Try it now. http://p.sf.net/sfu/bds _______________________________________________ Opensaf-users mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/opensaf-users
