Thanks. I resolved the issue by sync the imm.xml file on two controllers.  /Ted

-----Original Message-----
From: Anders Bjornerstedt [mailto:[email protected]] 
Sent: Monday, July 14, 2014 4:13 PM
To: Yao Cheng LIANG
Cc: [email protected]; santosh satapathy
Subject: Re: [users] One of the controller wait for sync

Hi ,

The sync request from SC2 clearly reaches SC1.
Is any sync started at SC1 ?
I cant see because the syslog snippet from SC1 is minimal, truncated right 
after the request arrives.

/Anders Bjornerstedt

Yao Cheng LIANG wrote:
> Dear all,
>
> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message 
> appears in /var/log/message on sc-1:
> ------------------------------------------------------
> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id  328207 : EE_ID 
> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1
> Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on 
> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: 
> IMMND on controller (not currently coord) requests sync Jul 12 
> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync 
> sync-pid:8930 epoch:0
> ----------------------------------------------------------------------
> --------------------------------------------------------
>
> while on sc-2, below message appears in /var/log/message:
> ----------------------------------------------------------------------
> --------------------------------------------------------
> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 
> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 WR20-64_32 
> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 
> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK Jul 12 
> 22:35:26 WR20-64_32 osafrded[8878]: Started
>
> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 
> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK Jul 
> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK Jul 12 
> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has 
> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: 
> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : 
> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd 
> startup - OK Jul 12 22:35:26 WR20-64_32 
> /etc/redhat-lsb/lsb_log_message:  - OK Jul 12 22:35:26 WR20-64_32 
> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: 
> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 
> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 
> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK Jul 12 
> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 
> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 
> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 
> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK 
> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message:  - OK Jul 
> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: 
> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 
> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING 
> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 
> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 
> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> 
> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: 
> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 
> osafimmnd[8930]: This node still waiting to be sync'ed after 20 
> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still 
> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 
> osafimmnd[8930]: This node still waiting to be sync'ed after 60 
> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still 
> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 
> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 WR20-64_32 
> osafimmnd[8930]: This node still waiting to be sync'ed after 100 
> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync 
> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 
> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 
> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: 
> This node still waiting to be sync'ed after 140 seconds Jul 12 
> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 
> detected at standby immd!! f2. Possible failover Jul 12 22:37:52 
> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp 
> to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message 
> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: 
> This node still waiting to be sync'ed after 160 seconds Jul 12 
> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be 
> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: 
> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: 
> This node still waiting to be sync'ed after 200 seconds Jul 12 
> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when 
> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add 
> New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 
> osafimmnd[8930]: This node still waiting to be sync'ed after 220 
> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still 
> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 
> osafimmnd[8930]: This node still waiting to be sync'ed after 260 
> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still 
> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 
> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 WR20-64_32 
> osafimmnd[8930]: This node still waiting to be sync'ed after 300 
> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync 
> request, when IMM_NODE_ISOLATED
> ----------------------------------------------------------------------
> ---------------------------------------------------------
>
> But I reverse the order - i.e. to start sc-2 and then sc-2, both 
> controller can be started successfully
>
> Could anyone tell me what's wrong?
>
> Thanks.
>
> Ted
>
>
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> This message (including any attachments) is for the named 
> addressee(s)'s use only. It may contain sensitive, confidential, 
> private proprietary or legally privileged information intended for a 
> specific individual and purpose, and is protected by law. If you are 
> not the intended recipient, please immediately delete it and all 
> copies of it from your system, destroy any hard copies of it and 
> notify the sender. Any use, disclosure, copying, or distribution of 
> this message and/or any attachments is strictly prohibited.
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>
> ----------------------------------------------------------------------
> -------- _______________________________________________
> Opensaf-users mailing list
> [email protected]
> https://lists.sourceforge.net/lists/listinfo/opensaf-users
>   



~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
This message (including any attachments) is for the named
addressee(s)'s use only. It may contain sensitive, confidential,
private proprietary or legally privileged information intended for a
specific individual and purpose, and is protected by law. If you are
not the intended recipient, please immediately delete it and all copies
of it from your system, destroy any hard copies of it
and notify the sender. Any use, disclosure, copying, or distribution of
this message and/or any attachments is strictly prohibited.
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~


------------------------------------------------------------------------------
Want fast and easy access to all the code in your enterprise? Index and
search up to 200,000 lines of code with a free copy of Black Duck&#174;
Code Sight&#153; - the same software that powers the world's largest code
search on Ohloh, the Black Duck Open Hub! Try it now.
http://p.sf.net/sfu/bds
_______________________________________________
Opensaf-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/opensaf-users

Reply via email to