From: santosh s. <san...@gm...> - 2014-07-11 06:42:36
|
Hi Again, I am getting the below health check issue while integrating the AIS APIs for one component as a sa-aware on the controller nodes in 2n mode. Upon starting/stopping the health check I get the below error. At the same time the CSI assignment and callbacks are all fine. Error: AisNotExistException: The entity does not exist. Health check for the component fails even though component registration is successful and CSI assignment for the ha state as ahctive. Even it takes further CSI assignment for locking and unlocking etc. The component name is same as registered and health check key as same value which is working for other components hosted in payload nodes with no-redundancy model. Healthchecktype is configured in the imm.xml as the demo application.xml. What else could be the issue here. Does the health check key needs to be different for different redundancy models? Any guess here? -- Thanks and Regards, Santosh |
From: Yao C. L. <yc...@as...> - 2014-07-12 14:47:10
|
Dear all, I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: ------------------------------------------------------ Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : EE_ID safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: IMMND on controller (not currently coord) requests sync Jul 12 22:35:26 localhost osafimmd[11706]: Node 5020f request sync sync-pid:8930 epoch:0 ------------------------------------------------------------------------------------------------------------------------------ while on sc-2, below message appears in /var/log/message: ------------------------------------------------------------------------------------------------------------------------------ Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd startup - OK Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 20 seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 60 seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 100 seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 140 seconds Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 160 seconds Jul 12 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 200 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 220 seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 260 seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 300 seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync request, when IMM_NODE_ISOLATED ------------------------------------------------------------------------------------------------------------------------------- But I reverse the order - i.e. to start sc-2 and then sc-2, both controller can be started successfully Could anyone tell me what’s wrong? Thanks. Ted ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Anders B. <and...@er...> - 2014-07-14 08:13:20
|
Hi , The sync request from SC2 clearly reaches SC1. Is any sync started at SC1 ? I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. /Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear all, > > I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: > ------------------------------------------------------ > Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : EE_ID safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on STANDBY Controller at 5020f > Jul 12 22:35:26 localhost osafimmd[11706]: IMMND on controller (not currently coord) requests sync > Jul 12 22:35:26 localhost osafimmd[11706]: Node 5020f request sync sync-pid:8930 epoch:0 > ------------------------------------------------------------------------------------------------------------------------------ > > while on sc-2, below message appears in /var/log/message: > ------------------------------------------------------------------------------------------------------------------------------ > Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services > Jul 12 22:35:26 WR20-64_32 osafdtmd[8860]: Started > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started > > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has active state => Standby role > Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: Started > Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Started > Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: Initialization Success, role STANDBY > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Started > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Initialization Success > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: Director Service is up > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING > Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: NODE STATE-> IMM_NODE_ISOLATED > Jul 12 22:35:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 20 seconds > Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 40 seconds > Jul 12 22:36:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 60 seconds > Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 80 seconds > Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 > Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 100 seconds > Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync request, when IMM_NODE_ISOLATED > Jul 12 22:37:16 WR20-64_32 osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 > Jul 12 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 120 seconds > Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 140 seconds > Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 detected at standby immd!! f2. Possible failover > Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp to peer IMMD > Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message count:1856 + 1 != 1856 > Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 160 seconds > Jul 12 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 180 seconds > Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 2000 > Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 200 seconds > Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when IMM_NODE_ISOLATED > Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add New incoming connection to fd : 22 > Jul 12 22:39:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 220 seconds > Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 240 seconds > Jul 12 22:39:46 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 260 seconds > Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 280 seconds > Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 > Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be sync'ed after 300 seconds > Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync request, when IMM_NODE_ISOLATED > ------------------------------------------------------------------------------------------------------------------------------- > > But I reverse the order - i.e. to start sc-2 and then sc-2, both controller can be started successfully > > Could anyone tell me what’s wrong? > > Thanks. > > Ted > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all copies > of it from your system, destroy any hard copies of it > and notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ------------------------------------------------------------------------------ > _______________________________________________ > Opensaf-users mailing list > Ope...@li... > https://lists.sourceforge.net/lists/listinfo/opensaf-users > |
From: Yao C. L. <yc...@as...> - 2014-07-14 08:32:35
|
Thanks. I resolved the issue by sync the imm.xml file on two controllers. /Ted -----Original Message----- From: Anders Bjornerstedt [mailto:and...@er...] Sent: Monday, July 14, 2014 4:13 PM To: Yao Cheng LIANG Cc: ope...@li...; santosh satapathy Subject: Re: [users] One of the controller wait for sync Hi , The sync request from SC2 clearly reaches SC1. Is any sync started at SC1 ? I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. /Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear all, > > I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: > ------------------------------------------------------ > Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : EE_ID > safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on > STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: > IMMND on controller (not currently coord) requests sync Jul 12 > 22:35:26 localhost osafimmd[11706]: Node 5020f request sync > sync-pid:8930 epoch:0 > ---------------------------------------------------------------------- > -------------------------------------------------------- > > while on sc-2, below message appears in /var/log/message: > ---------------------------------------------------------------------- > -------------------------------------------------------- > Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 > 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 WR20-64_32 > /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 > 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 > 22:35:26 WR20-64_32 osafrded[8878]: Started > > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 > WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK Jul > 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 > 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has > active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: > Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : > safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd > startup - OK Jul 12 22:35:26 WR20-64_32 > /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 > osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: > Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 > /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 > 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 > 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 > WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 > WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 > WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul > 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: > IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 > WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING > --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 > osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 > osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> > IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: > NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 20 > seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 60 > seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 > osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 100 > seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync > request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 > osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 > 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: > This node still waiting to be sync'ed after 140 seconds Jul 12 > 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 > detected at standby immd!! f2. Possible failover Jul 12 22:37:52 > WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp > to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message > count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: > This node still waiting to be sync'ed after 160 seconds Jul 12 > 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: > REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: > This node still waiting to be sync'ed after 200 seconds Jul 12 > 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when > IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add > New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 220 > seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 260 > seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 > osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 300 > seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync > request, when IMM_NODE_ISOLATED > ---------------------------------------------------------------------- > --------------------------------------------------------- > > But I reverse the order - i.e. to start sc-2 and then sc-2, both > controller can be started successfully > > Could anyone tell me what's wrong? > > Thanks. > > Ted > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all > copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ---------------------------------------------------------------------- > -------- _______________________________________________ > Opensaf-users mailing list > Ope...@li... > https://lists.sourceforge.net/lists/listinfo/opensaf-users > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Anders B. <and...@er...> - 2014-07-14 08:47:50
|
Hi, There is no such thing as "sync the imm.xml file". Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. An imm.cml file can be used for loading. Sync is performaed by nodes that *missed* loading. (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). /Anders Bjornerstedt Yao Cheng LIANG wrote: > Thanks. I resolved the issue by sync the imm.xml file on two controllers. /Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:13 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi , > > The sync request from SC2 clearly reaches SC1. > Is any sync started at SC1 ? > I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Dear all, >> >> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >> ------------------------------------------------------ >> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : EE_ID >> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 >> Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >> IMMND on controller (not currently coord) requests sync Jul 12 >> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >> sync-pid:8930 epoch:0 >> ---------------------------------------------------------------------- >> -------------------------------------------------------- >> >> while on sc-2, below message appears in /var/log/message: >> ---------------------------------------------------------------------- >> -------------------------------------------------------- >> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 >> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 WR20-64_32 >> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >> 22:35:26 WR20-64_32 osafrded[8878]: Started >> >> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK Jul >> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 >> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd >> startup - OK Jul 12 22:35:26 WR20-64_32 >> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul >> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >> This node still waiting to be sync'ed after 140 seconds Jul 12 >> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 >> detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp >> to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >> This node still waiting to be sync'ed after 160 seconds Jul 12 >> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >> This node still waiting to be sync'ed after 200 seconds Jul 12 >> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add >> New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >> request, when IMM_NODE_ISOLATED >> ---------------------------------------------------------------------- >> --------------------------------------------------------- >> >> But I reverse the order - i.e. to start sc-2 and then sc-2, both >> controller can be started successfully >> >> Could anyone tell me what's wrong? >> >> Thanks. >> >> Ted >> >> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> ---------------------------------------------------------------------- >> -------- _______________________________________________ >> Opensaf-users mailing list >> Ope...@li... >> https://lists.sourceforge.net/lists/listinfo/opensaf-users >> >> > > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all copies > of it from your system, destroy any hard copies of it > and notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > |
From: Yao C. L. <yc...@as...> - 2014-07-14 08:49:32
|
Dear Anders, Thanks for clarification. Here "sync" I mean cop the file from one node to the other. Ted -----Original Message----- From: Anders Bjornerstedt [mailto:and...@er...] Sent: Monday, July 14, 2014 4:48 PM To: Yao Cheng LIANG Cc: ope...@li...; santosh satapathy Subject: Re: [users] One of the controller wait for sync Hi, There is no such thing as "sync the imm.xml file". Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. An imm.cml file can be used for loading. Sync is performaed by nodes that *missed* loading. (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). /Anders Bjornerstedt Yao Cheng LIANG wrote: > Thanks. I resolved the issue by sync the imm.xml file on two > controllers. /Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:13 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi , > > The sync request from SC2 clearly reaches SC1. > Is any sync started at SC1 ? > I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Dear all, >> >> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >> ------------------------------------------------------ >> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : >> EE_ID >> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 >> Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >> IMMND on controller (not currently coord) requests sync Jul 12 >> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >> sync-pid:8930 epoch:0 >> --------------------------------------------------------------------- >> - >> -------------------------------------------------------- >> >> while on sc-2, below message appears in /var/log/message: >> --------------------------------------------------------------------- >> - >> -------------------------------------------------------- >> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 >> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 >> WR20-64_32 >> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >> 22:35:26 WR20-64_32 osafrded[8878]: Started >> >> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK >> Jul >> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 >> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd >> startup - OK Jul 12 22:35:26 WR20-64_32 >> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul >> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 >> WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >> This node still waiting to be sync'ed after 140 seconds Jul 12 >> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller >> f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp >> to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >> This node still waiting to be sync'ed after 160 seconds Jul 12 >> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >> This node still waiting to be sync'ed after 200 seconds Jul 12 >> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add >> New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 >> WR20-64_32 >> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >> request, when IMM_NODE_ISOLATED >> --------------------------------------------------------------------- >> - >> --------------------------------------------------------- >> >> But I reverse the order - i.e. to start sc-2 and then sc-2, both >> controller can be started successfully >> >> Could anyone tell me what's wrong? >> >> Thanks. >> >> Ted >> >> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> --------------------------------------------------------------------- >> - >> -------- _______________________________________________ >> Opensaf-users mailing list >> Ope...@li... >> https://lists.sourceforge.net/lists/listinfo/opensaf-users >> >> > > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all > copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Anders B. <and...@er...> - 2014-07-14 08:56:22
|
Hi, Sounds like you dont have a shared file system mounted between SC1 and SC2. That means you can not run what is called 1PBE which relies on a shared filesystem. But you van run 0PBE or 2PBE. PBE = Persistent Back End. But in any case, your initial problem of SC2 not getting synced is strange. If you have not already done so, you need to read the documentation for the IMM. Either the OpenSAF_IMMSV_PR.doc or the osaf/services/saf/immsv/README. In particular the overview parts that explain imm loading, imm sync and PBE. /Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear Anders, > > Thanks for clarification. Here "sync" I mean cop the file from one node to the other. > > Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:48 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi, > > There is no such thing as "sync the imm.xml file". > Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. > > An imm.cml file can be used for loading. > Sync is performaed by nodes that *missed* loading. > > (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Thanks. I resolved the issue by sync the imm.xml file on two >> controllers. /Ted >> >> -----Original Message----- >> From: Anders Bjornerstedt [mailto:and...@er...] >> Sent: Monday, July 14, 2014 4:13 PM >> To: Yao Cheng LIANG >> Cc: ope...@li...; santosh satapathy >> Subject: Re: [users] One of the controller wait for sync >> >> Hi , >> >> The sync request from SC2 clearly reaches SC1. >> Is any sync started at SC1 ? >> I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. >> >> /Anders Bjornerstedt >> >> Yao Cheng LIANG wrote: >> >> >>> Dear all, >>> >>> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >>> ------------------------------------------------------ >>> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : >>> EE_ID >>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 >>> Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >>> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >>> IMMND on controller (not currently coord) requests sync Jul 12 >>> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >>> sync-pid:8930 epoch:0 >>> --------------------------------------------------------------------- >>> - >>> -------------------------------------------------------- >>> >>> while on sc-2, below message appears in /var/log/message: >>> --------------------------------------------------------------------- >>> - >>> -------------------------------------------------------- >>> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 >>> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 >>> WR20-64_32 >>> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafrded[8878]: Started >>> >>> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK >>> Jul >>> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >>> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >>> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 >>> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd >>> startup - OK Jul 12 22:35:26 WR20-64_32 >>> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >>> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >>> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >>> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >>> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul >>> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >>> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >>> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >>> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >>> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >>> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >>> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >>> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 >>> WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >>> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >>> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >>> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >>> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 140 seconds Jul 12 >>> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller >>> f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >>> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp >>> to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >>> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 160 seconds Jul 12 >>> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 200 seconds Jul 12 >>> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >>> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add >>> New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >>> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >>> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 >>> WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >>> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >>> request, when IMM_NODE_ISOLATED >>> --------------------------------------------------------------------- >>> - >>> --------------------------------------------------------- >>> >>> But I reverse the order - i.e. to start sc-2 and then sc-2, both >>> controller can be started successfully >>> >>> Could anyone tell me what's wrong? >>> >>> Thanks. >>> >>> Ted >>> >>> >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> This message (including any attachments) is for the named >>> addressee(s)'s use only. It may contain sensitive, confidential, >>> private proprietary or legally privileged information intended for a >>> specific individual and purpose, and is protected by law. If you are >>> not the intended recipient, please immediately delete it and all >>> copies of it from your system, destroy any hard copies of it and >>> notify the sender. Any use, disclosure, copying, or distribution of >>> this message and/or any attachments is strictly prohibited. >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> >>> --------------------------------------------------------------------- >>> - >>> -------- _______________________________________________ >>> Opensaf-users mailing list >>> Ope...@li... >>> https://lists.sourceforge.net/lists/listinfo/opensaf-users >>> >>> >>> >> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> >> > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all copies > of it from your system, destroy any hard copies of it > and notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > |
From: Yao C. L. <yc...@as...> - 2014-07-14 09:02:19
|
Thanks. I have read SAFForum spec on imm, but have not read the one in opensaf. I just skimmed it, and will read in depth sonn. Thanks for advice. Ted -----Original Message----- From: Anders Bjornerstedt [mailto:and...@er...] Sent: Monday, July 14, 2014 4:56 PM To: Yao Cheng LIANG Cc: ope...@li...; santosh satapathy Subject: Re: [users] One of the controller wait for sync Hi, Sounds like you dont have a shared file system mounted between SC1 and SC2. That means you can not run what is called 1PBE which relies on a shared filesystem. But you van run 0PBE or 2PBE. PBE = Persistent Back End. But in any case, your initial problem of SC2 not getting synced is strange. If you have not already done so, you need to read the documentation for the IMM. Either the OpenSAF_IMMSV_PR.doc or the osaf/services/saf/immsv/README. In particular the overview parts that explain imm loading, imm sync and PBE. /Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear Anders, > > Thanks for clarification. Here "sync" I mean cop the file from one node to the other. > > Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:48 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi, > > There is no such thing as "sync the imm.xml file". > Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. > > An imm.cml file can be used for loading. > Sync is performaed by nodes that *missed* loading. > > (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Thanks. I resolved the issue by sync the imm.xml file on two >> controllers. /Ted >> >> -----Original Message----- >> From: Anders Bjornerstedt [mailto:and...@er...] >> Sent: Monday, July 14, 2014 4:13 PM >> To: Yao Cheng LIANG >> Cc: ope...@li...; santosh satapathy >> Subject: Re: [users] One of the controller wait for sync >> >> Hi , >> >> The sync request from SC2 clearly reaches SC1. >> Is any sync started at SC1 ? >> I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. >> >> /Anders Bjornerstedt >> >> Yao Cheng LIANG wrote: >> >> >>> Dear all, >>> >>> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >>> ------------------------------------------------------ >>> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : >>> EE_ID >>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>> 1 Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >>> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >>> IMMND on controller (not currently coord) requests sync Jul 12 >>> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >>> sync-pid:8930 epoch:0 >>> -------------------------------------------------------------------- >>> - >>> - >>> -------------------------------------------------------- >>> >>> while on sc-2, below message appears in /var/log/message: >>> -------------------------------------------------------------------- >>> - >>> - >>> -------------------------------------------------------- >>> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul >>> 12 >>> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 >>> WR20-64_32 >>> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafrded[8878]: Started >>> >>> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK >>> Jul >>> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >>> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >>> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>> 1 Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: >>> osaffmd startup - OK Jul 12 22:35:26 WR20-64_32 >>> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >>> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >>> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >>> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >>> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK >>> Jul >>> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >>> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >>> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >>> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >>> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >>> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >>> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >>> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 >>> WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >>> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >>> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >>> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >>> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 140 seconds Jul 12 >>> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller >>> f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >>> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not >>> mbcp to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >>> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 160 seconds Jul 12 >>> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 200 seconds Jul 12 >>> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >>> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: >>> add New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >>> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >>> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 >>> WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >>> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >>> request, when IMM_NODE_ISOLATED >>> -------------------------------------------------------------------- >>> - >>> - >>> --------------------------------------------------------- >>> >>> But I reverse the order - i.e. to start sc-2 and then sc-2, both >>> controller can be started successfully >>> >>> Could anyone tell me what's wrong? >>> >>> Thanks. >>> >>> Ted >>> >>> >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> This message (including any attachments) is for the named >>> addressee(s)'s use only. It may contain sensitive, confidential, >>> private proprietary or legally privileged information intended for a >>> specific individual and purpose, and is protected by law. If you are >>> not the intended recipient, please immediately delete it and all >>> copies of it from your system, destroy any hard copies of it and >>> notify the sender. Any use, disclosure, copying, or distribution of >>> this message and/or any attachments is strictly prohibited. >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> >>> -------------------------------------------------------------------- >>> - >>> - >>> -------- _______________________________________________ >>> Opensaf-users mailing list >>> Ope...@li... >>> https://lists.sourceforge.net/lists/listinfo/opensaf-users >>> >>> >>> >> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> >> > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all > copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Yao C. L. <yc...@as...> - 2014-07-15 01:51:58
|
The issue came up again. Could anyone tell me how to identify where is wrong? I am using OpenSAF 4.2.0. Below is what is seen from the node requesting sync.: ----------------------------------------------------- Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Started Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Initialization Success Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Director Service is up Jul 15 09:42:52 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK Jul 15 09:42:52 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: REQUESTING SYNC Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: NODE STATE-> IMM_NODE_ISOLATED Jul 15 09:42:57 WR20-64_32 osafdtmd[30313]: DTM: dtm_node_add failed .node_ip : 192.168.211.181 Jul 15 09:43:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 20 seconds Jul 15 09:43:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 40 seconds Jul 15 09:43:53 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 60 seconds Jul 15 09:44:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 80 seconds Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: REQUESTING SYNC AGAIN 1000 Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 100 seconds Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: Redundant sync request, when IMM_NODE_ISOLATED Jul 15 09:44:53 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 120 seconds Jul 15 09:45:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 140 seconds Jul 15 09:45:26 WR20-64_32 syslog-ng[3560]: STATS: dropped 0 Jul 15 09:45:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 160 seconds And below is from another SC: ---------------------------------------------------------------- Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Implementer connected: 10 (MsgQueueService133135) <0, 2080f> Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safSISU=safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=CPND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=CPND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=GLND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=GLND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=MQND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=MQND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=IMMND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=IMMND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=SMFND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=SMFND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=AMFWDOG\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=AMFWDOG,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safSISU=safSu=PL-5\#safSg=DpAmfFGW\#safApp=DpAmfFGWType,safSi=DpAmfFGW,safApp=DpAmfFGWType' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=DpAmfFGW\#safSu=PL-5\#safSg=DpAmfFGW\#safApp=DpAmfFGWType,safCsi=DpAmfFGW,safSi=DpAmfFGW,safApp=DpAmfFGWType' by Impl id: 3 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCkpt=fgw,safApp=safCkptService' by Impl id: 8 Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safReplica=safNode=PL-5\#safCluster=myClmCluster,safCkpt=fgw,safApp=safCkptService' by Impl id: 8 Jul 16 01:42:52 WR20-64_25 osafdtmd[22925]: DTM: add New incoming connection to fd : 77 Jul 16 01:42:52 WR20-64_25 osafdtmd[22925]: DTM: add New incoming connection to fd : 78 Jul 16 01:42:53 WR20-64_25 osaffmd[22965]: Peer Node_id 328207 : EE_ID safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: New IMMND process is on ACTIVE Controller at 2030f Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: New IMMND process is on STANDBY Controller at 5020f Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 Thanks. Ted -----Original Message----- From: Anders Bjornerstedt [mailto:and...@er...] Sent: Monday, July 14, 2014 4:56 PM To: Yao Cheng LIANG Cc: ope...@li...; santosh satapathy Subject: Re: [users] One of the controller wait for sync Hi, Sounds like you dont have a shared file system mounted between SC1 and SC2. That means you can not run what is called 1PBE which relies on a shared filesystem. But you van run 0PBE or 2PBE. PBE = Persistent Back End. But in any case, your initial problem of SC2 not getting synced is strange. If you have not already done so, you need to read the documentation for the IMM. Either the OpenSAF_IMMSV_PR.doc or the osaf/services/saf/immsv/README. In particular the overview parts that explain imm loading, imm sync and PBE. /Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear Anders, > > Thanks for clarification. Here "sync" I mean cop the file from one node to the other. > > Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:48 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi, > > There is no such thing as "sync the imm.xml file". > Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. > > An imm.cml file can be used for loading. > Sync is performaed by nodes that *missed* loading. > > (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Thanks. I resolved the issue by sync the imm.xml file on two >> controllers. /Ted >> >> -----Original Message----- >> From: Anders Bjornerstedt [mailto:and...@er...] >> Sent: Monday, July 14, 2014 4:13 PM >> To: Yao Cheng LIANG >> Cc: ope...@li...; santosh satapathy >> Subject: Re: [users] One of the controller wait for sync >> >> Hi , >> >> The sync request from SC2 clearly reaches SC1. >> Is any sync started at SC1 ? >> I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. >> >> /Anders Bjornerstedt >> >> Yao Cheng LIANG wrote: >> >> >>> Dear all, >>> >>> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >>> ------------------------------------------------------ >>> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : >>> EE_ID >>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>> 1 Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >>> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >>> IMMND on controller (not currently coord) requests sync Jul 12 >>> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >>> sync-pid:8930 epoch:0 >>> -------------------------------------------------------------------- >>> - >>> - >>> -------------------------------------------------------- >>> >>> while on sc-2, below message appears in /var/log/message: >>> -------------------------------------------------------------------- >>> - >>> - >>> -------------------------------------------------------- >>> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul >>> 12 >>> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 >>> WR20-64_32 >>> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafrded[8878]: Started >>> >>> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK >>> Jul >>> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >>> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >>> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>> 1 Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: >>> osaffmd startup - OK Jul 12 22:35:26 WR20-64_32 >>> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >>> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >>> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >>> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >>> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK >>> Jul >>> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >>> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >>> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >>> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >>> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >>> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >>> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >>> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >>> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 >>> WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >>> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >>> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >>> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >>> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 140 seconds Jul 12 >>> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller >>> f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >>> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not >>> mbcp to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >>> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 160 seconds Jul 12 >>> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>> This node still waiting to be sync'ed after 200 seconds Jul 12 >>> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >>> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: >>> add New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >>> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >>> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >>> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >>> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 >>> WR20-64_32 >>> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >>> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >>> request, when IMM_NODE_ISOLATED >>> -------------------------------------------------------------------- >>> - >>> - >>> --------------------------------------------------------- >>> >>> But I reverse the order - i.e. to start sc-2 and then sc-2, both >>> controller can be started successfully >>> >>> Could anyone tell me what's wrong? >>> >>> Thanks. >>> >>> Ted >>> >>> >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> This message (including any attachments) is for the named >>> addressee(s)'s use only. It may contain sensitive, confidential, >>> private proprietary or legally privileged information intended for a >>> specific individual and purpose, and is protected by law. If you are >>> not the intended recipient, please immediately delete it and all >>> copies of it from your system, destroy any hard copies of it and >>> notify the sender. Any use, disclosure, copying, or distribution of >>> this message and/or any attachments is strictly prohibited. >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> >>> -------------------------------------------------------------------- >>> - >>> - >>> -------- _______________________________________________ >>> Opensaf-users mailing list >>> Ope...@li... >>> https://lists.sourceforge.net/lists/listinfo/opensaf-users >>> >>> >>> >> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> >> > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all > copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Anders B. <and...@er...> - 2014-07-15 07:12:58
|
There seems to be some problem with how you have configured or built the system. I can not say what the problem is, but one hint is: Jul 15 09:42:57 WR20-64_32 osafdtmd[30313]: DTM: dtm_node_add failed .node_ip : 192.168.211.181 Have you read and followed the instructions of the README in the top directory ? /Anders Bjornerstedt Yao Cheng LIANG wrote: > The issue came up again. Could anyone tell me how to identify where is wrong? I am using OpenSAF 4.2.0. > > Below is what is seen from the node requesting sync.: > ----------------------------------------------------- > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Started > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Initialization Success > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Director Service is up > Jul 15 09:42:52 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK > Jul 15 09:42:52 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING > Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: REQUESTING SYNC > Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING > Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: NODE STATE-> IMM_NODE_ISOLATED > Jul 15 09:42:57 WR20-64_32 osafdtmd[30313]: DTM: dtm_node_add failed .node_ip : 192.168.211.181 > Jul 15 09:43:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 20 seconds > Jul 15 09:43:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 40 seconds > Jul 15 09:43:53 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 60 seconds > Jul 15 09:44:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 80 seconds > Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: REQUESTING SYNC AGAIN 1000 > Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 100 seconds > Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: Redundant sync request, when IMM_NODE_ISOLATED > Jul 15 09:44:53 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 120 seconds > Jul 15 09:45:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 140 seconds > Jul 15 09:45:26 WR20-64_32 syslog-ng[3560]: STATS: dropped 0 > Jul 15 09:45:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 160 seconds > > And below is from another SC: > ---------------------------------------------------------------- > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Implementer connected: 10 (MsgQueueService133135) <0, 2080f> > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safSISU=safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=CPND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=CPND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=GLND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=GLND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=MQND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=MQND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=IMMND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=IMMND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=SMFND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=SMFND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=AMFWDOG\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=AMFWDOG,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safSISU=safSu=PL-5\#safSg=DpAmfFGW\#safApp=DpAmfFGWType,safSi=DpAmfFGW,safApp=DpAmfFGWType' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=DpAmfFGW\#safSu=PL-5\#safSg=DpAmfFGW\#safApp=DpAmfFGWType,safCsi=DpAmfFGW,safSi=DpAmfFGW,safApp=DpAmfFGWType' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCkpt=fgw,safApp=safCkptService' by Impl id: 8 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safReplica=safNode=PL-5\#safCluster=myClmCluster,safCkpt=fgw,safApp=safCkptService' by Impl id: 8 > Jul 16 01:42:52 WR20-64_25 osafdtmd[22925]: DTM: add New incoming connection to fd : 77 > Jul 16 01:42:52 WR20-64_25 osafdtmd[22925]: DTM: add New incoming connection to fd : 78 > Jul 16 01:42:53 WR20-64_25 osaffmd[22965]: Peer Node_id 328207 : EE_ID safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: New IMMND process is on ACTIVE Controller at 2030f > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: New IMMND process is on STANDBY Controller at 5020f > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > > Thanks. > > Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:56 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi, > > Sounds like you dont have a shared file system mounted between SC1 and SC2. > That means you can not run what is called 1PBE which relies on a shared filesystem. > But you van run 0PBE or 2PBE. > PBE = Persistent Back End. > > But in any case, your initial problem of SC2 not getting synced is strange. > > If you have not already done so, you need to read the documentation for the IMM. > Either the OpenSAF_IMMSV_PR.doc or the osaf/services/saf/immsv/README. > In particular the overview parts that explain imm loading, imm sync and PBE. > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Dear Anders, >> >> Thanks for clarification. Here "sync" I mean cop the file from one node to the other. >> >> Ted >> >> -----Original Message----- >> From: Anders Bjornerstedt [mailto:and...@er...] >> Sent: Monday, July 14, 2014 4:48 PM >> To: Yao Cheng LIANG >> Cc: ope...@li...; santosh satapathy >> Subject: Re: [users] One of the controller wait for sync >> >> Hi, >> >> There is no such thing as "sync the imm.xml file". >> Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. >> >> An imm.cml file can be used for loading. >> Sync is performaed by nodes that *missed* loading. >> >> (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). >> >> /Anders Bjornerstedt >> >> Yao Cheng LIANG wrote: >> >> >>> Thanks. I resolved the issue by sync the imm.xml file on two >>> controllers. /Ted >>> >>> -----Original Message----- >>> From: Anders Bjornerstedt [mailto:and...@er...] >>> Sent: Monday, July 14, 2014 4:13 PM >>> To: Yao Cheng LIANG >>> Cc: ope...@li...; santosh satapathy >>> Subject: Re: [users] One of the controller wait for sync >>> >>> Hi , >>> >>> The sync request from SC2 clearly reaches SC1. >>> Is any sync started at SC1 ? >>> I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. >>> >>> /Anders Bjornerstedt >>> >>> Yao Cheng LIANG wrote: >>> >>> >>> >>>> Dear all, >>>> >>>> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >>>> ------------------------------------------------------ >>>> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : >>>> EE_ID >>>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>>> 1 Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >>>> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >>>> IMMND on controller (not currently coord) requests sync Jul 12 >>>> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >>>> sync-pid:8930 epoch:0 >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> -------------------------------------------------------- >>>> >>>> while on sc-2, below message appears in /var/log/message: >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> -------------------------------------------------------- >>>> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul >>>> 12 >>>> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 >>>> WR20-64_32 >>>> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >>>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>>> 22:35:26 WR20-64_32 osafrded[8878]: Started >>>> >>>> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >>>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK >>>> Jul >>>> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>>> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >>>> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >>>> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >>>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>>> 1 Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: >>>> osaffmd startup - OK Jul 12 22:35:26 WR20-64_32 >>>> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >>>> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >>>> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >>>> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >>>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>>> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >>>> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >>>> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >>>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >>>> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK >>>> Jul >>>> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >>>> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >>>> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >>>> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >>>> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >>>> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >>>> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >>>> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >>>> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >>>> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >>>> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 >>>> WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >>>> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >>>> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >>>> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >>>> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>>> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >>>> This node still waiting to be sync'ed after 140 seconds Jul 12 >>>> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller >>>> f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >>>> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not >>>> mbcp to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >>>> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >>>> This node still waiting to be sync'ed after 160 seconds Jul 12 >>>> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>>> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>>> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>>> This node still waiting to be sync'ed after 200 seconds Jul 12 >>>> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >>>> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: >>>> add New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >>>> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >>>> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >>>> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 >>>> WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >>>> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >>>> request, when IMM_NODE_ISOLATED >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> --------------------------------------------------------- >>>> >>>> But I reverse the order - i.e. to start sc-2 and then sc-2, both >>>> controller can be started successfully >>>> >>>> Could anyone tell me what's wrong? >>>> >>>> Thanks. >>>> >>>> Ted >>>> >>>> >>>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>>> This message (including any attachments) is for the named >>>> addressee(s)'s use only. It may contain sensitive, confidential, >>>> private proprietary or legally privileged information intended for a >>>> specific individual and purpose, and is protected by law. If you are >>>> not the intended recipient, please immediately delete it and all >>>> copies of it from your system, destroy any hard copies of it and >>>> notify the sender. Any use, disclosure, copying, or distribution of >>>> this message and/or any attachments is strictly prohibited. >>>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>>> >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> -------- _______________________________________________ >>>> Opensaf-users mailing list >>>> Ope...@li... >>>> https://lists.sourceforge.net/lists/listinfo/opensaf-users >>>> >>>> >>>> >>>> >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> This message (including any attachments) is for the named >>> addressee(s)'s use only. It may contain sensitive, confidential, >>> private proprietary or legally privileged information intended for a >>> specific individual and purpose, and is protected by law. If you are >>> not the intended recipient, please immediately delete it and all >>> copies of it from your system, destroy any hard copies of it and >>> notify the sender. Any use, disclosure, copying, or distribution of >>> this message and/or any attachments is strictly prohibited. >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> >>> >>> >>> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> >> > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all copies > of it from your system, destroy any hard copies of it > and notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > |
From: Yao C. L. <yc...@as...> - 2014-07-15 11:25:44
|
Dear Anders, The issue was resolved, the reason is that one of the PL node has configured the same slot_id with one of the SC. But I am just curiously about, how this has caused this issue? And I have found nowhere the use of slot_id except when calculate tipc address, but I am using tp. Thanks. Ted Sent from Windows Mail From: Anders Bjornerstedt Sent: Tuesday, July 15, 2014 3:13 PM To: Yao Cheng LIANG Cc: ope...@li..., santosh satapathy There seems to be some problem with how you have configured or built the system. I can not say what the problem is, but one hint is: Jul 15 09:42:57 WR20-64_32 osafdtmd[30313]: DTM: dtm_node_add failed .node_ip : 192.168.211.181 Have you read and followed the instructions of the README in the top directory ? /Anders Bjornerstedt Yao Cheng LIANG wrote: > The issue came up again. Could anyone tell me how to identify where is wrong? I am using OpenSAF 4.2.0. > > Below is what is seen from the node requesting sync.: > ----------------------------------------------------- > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Started > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Initialization Success > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: Director Service is up > Jul 15 09:42:52 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK > Jul 15 09:42:52 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING > Jul 15 09:42:52 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING --> IMM_SERVER_LOADING_PENDING > Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: REQUESTING SYNC > Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> IMM_SERVER_SYNC_PENDING > Jul 15 09:42:53 WR20-64_32 osafimmnd[30384]: NODE STATE-> IMM_NODE_ISOLATED > Jul 15 09:42:57 WR20-64_32 osafdtmd[30313]: DTM: dtm_node_add failed .node_ip : 192.168.211.181 > Jul 15 09:43:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 20 seconds > Jul 15 09:43:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 40 seconds > Jul 15 09:43:53 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 60 seconds > Jul 15 09:44:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 80 seconds > Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: REQUESTING SYNC AGAIN 1000 > Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 100 seconds > Jul 15 09:44:33 WR20-64_32 osafimmnd[30384]: Redundant sync request, when IMM_NODE_ISOLATED > Jul 15 09:44:53 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 120 seconds > Jul 15 09:45:13 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 140 seconds > Jul 15 09:45:26 WR20-64_32 syslog-ng[3560]: STATS: dropped 0 > Jul 15 09:45:33 WR20-64_32 osafimmnd[30384]: This node still waiting to be sync'ed after 160 seconds > > And below is from another SC: > ---------------------------------------------------------------- > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Implementer connected: 10 (MsgQueueService133135) <0, 2080f> > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safSISU=safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=CPND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=CPND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=GLND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=GLND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=MQND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=MQND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=IMMND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=IMMND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=SMFND\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=SMFND,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=AMFWDOG\#safSu=PL-5\#safSg=NoRed\#safApp=OpenSAF,safCsi=AMFWDOG,safSi=NoRed1,safApp=OpenSAF' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safSISU=safSu=PL-5\#safSg=DpAmfFGW\#safApp=DpAmfFGWType,safSi=DpAmfFGW,safApp=DpAmfFGWType' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCSIComp=safComp=DpAmfFGW\#safSu=PL-5\#safSg=DpAmfFGW\#safApp=DpAmfFGWType,safCsi=DpAmfFGW,safSi=DpAmfFGW,safApp=DpAmfFGWType' by Impl id: 3 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safCkpt=fgw,safApp=safCkptService' by Impl id: 8 > Jul 16 01:42:14 WR20-64_25 osafimmnd[22998]: Create runtime object 'safReplica=safNode=PL-5\#safCluster=myClmCluster,safCkpt=fgw,safApp=safCkptService' by Impl id: 8 > Jul 16 01:42:52 WR20-64_25 osafdtmd[22925]: DTM: add New incoming connection to fd : 77 > Jul 16 01:42:52 WR20-64_25 osafdtmd[22925]: DTM: add New incoming connection to fd : 78 > Jul 16 01:42:53 WR20-64_25 osaffmd[22965]: Peer Node_id 328207 : EE_ID safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: New IMMND process is on ACTIVE Controller at 2030f > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: New IMMND process is on STANDBY Controller at 5020f > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:42:53 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:44:33 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:46:13 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: Node 2030f request sync sync-pid:940 epoch:0 > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: IMMND on controller (not currently coord) requests sync > Jul 16 01:47:53 WR20-64_25 osafimmd[22981]: Node 5020f request sync sync-pid:30384 epoch:0 > > Thanks. > > Ted > > -----Original Message----- > From: Anders Bjornerstedt [mailto:and...@er...] > Sent: Monday, July 14, 2014 4:56 PM > To: Yao Cheng LIANG > Cc: ope...@li...; santosh satapathy > Subject: Re: [users] One of the controller wait for sync > > Hi, > > Sounds like you dont have a shared file system mounted between SC1 and SC2. > That means you can not run what is called 1PBE which relies on a shared filesystem. > But you van run 0PBE or 2PBE. > PBE = Persistent Back End. > > But in any case, your initial problem of SC2 not getting synced is strange. > > If you have not already done so, you need to read the documentation for the IMM. > Either the OpenSAF_IMMSV_PR.doc or the osaf/services/saf/immsv/README. > In particular the overview parts that explain imm loading, imm sync and PBE. > > /Anders Bjornerstedt > > Yao Cheng LIANG wrote: > >> Dear Anders, >> >> Thanks for clarification. Here "sync" I mean cop the file from one node to the other. >> >> Ted >> >> -----Original Message----- >> From: Anders Bjornerstedt [mailto:and...@er...] >> Sent: Monday, July 14, 2014 4:48 PM >> To: Yao Cheng LIANG >> Cc: ope...@li...; santosh satapathy >> Subject: Re: [users] One of the controller wait for sync >> >> Hi, >> >> There is no such thing as "sync the imm.xml file". >> Sync is a protocol where the IMMND at one of the SCs broadcastrs the imm contents (from memory) to any nodes that are "empty" and ready to receive the sync/data. Any node that has sent a sync request is ready to receive the sync. >> >> An imm.cml file can be used for loading. >> Sync is performaed by nodes that *missed* loading. >> >> (An imm.xml file can also be used to create a ccb using 'immcfg -f' but I dont think that is what you meant). >> >> /Anders Bjornerstedt >> >> Yao Cheng LIANG wrote: >> >> >>> Thanks. I resolved the issue by sync the imm.xml file on two >>> controllers. /Ted >>> >>> -----Original Message----- >>> From: Anders Bjornerstedt [mailto:and...@er...] >>> Sent: Monday, July 14, 2014 4:13 PM >>> To: Yao Cheng LIANG >>> Cc: ope...@li...; santosh satapathy >>> Subject: Re: [users] One of the controller wait for sync >>> >>> Hi , >>> >>> The sync request from SC2 clearly reaches SC1. >>> Is any sync started at SC1 ? >>> I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. >>> >>> /Anders Bjornerstedt >>> >>> Yao Cheng LIANG wrote: >>> >>> >>> >>>> Dear all, >>>> >>>> I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: >>>> ------------------------------------------------------ >>>> Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : >>>> EE_ID >>>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>>> 1 Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on >>>> STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: >>>> IMMND on controller (not currently coord) requests sync Jul 12 >>>> 22:35:26 localhost osafimmd[11706]: Node 5020f request sync >>>> sync-pid:8930 epoch:0 >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> -------------------------------------------------------- >>>> >>>> while on sc-2, below message appears in /var/log/message: >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> -------------------------------------------------------- >>>> Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul >>>> 12 >>>> 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 >>>> WR20-64_32 >>>> /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 >>>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>>> 22:35:26 WR20-64_32 osafrded[8878]: Started >>>> >>>> Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 >>>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK >>>> Jul >>>> 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>>> 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has >>>> active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: >>>> Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : >>>> safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_ >>>> 1 Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: >>>> osaffmd startup - OK Jul 12 22:35:26 WR20-64_32 >>>> /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 >>>> osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: >>>> Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 >>>> /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 >>>> 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 >>>> 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 >>>> WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 >>>> WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 >>>> WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK >>>> Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK >>>> Jul >>>> 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: >>>> IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 >>>> WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING >>>> --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 >>>> osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 >>>> osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> >>>> IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: >>>> NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 20 >>>> seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 60 >>>> seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 >>>> osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 >>>> WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 100 >>>> seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync >>>> request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 >>>> osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 >>>> 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>>> sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: >>>> This node still waiting to be sync'ed after 140 seconds Jul 12 >>>> 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller >>>> f3 detected at standby immd!! f2. Possible failover Jul 12 22:37:52 >>>> WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not >>>> mbcp to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message >>>> count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: >>>> This node still waiting to be sync'ed after 160 seconds Jul 12 >>>> 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be >>>> sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>>> REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: >>>> This node still waiting to be sync'ed after 200 seconds Jul 12 >>>> 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when >>>> IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: >>>> add New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 220 >>>> seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 260 >>>> seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still >>>> waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 >>>> osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 >>>> WR20-64_32 >>>> osafimmnd[8930]: This node still waiting to be sync'ed after 300 >>>> seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync >>>> request, when IMM_NODE_ISOLATED >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> --------------------------------------------------------- >>>> >>>> But I reverse the order - i.e. to start sc-2 and then sc-2, both >>>> controller can be started successfully >>>> >>>> Could anyone tell me what's wrong? >>>> >>>> Thanks. >>>> >>>> Ted >>>> >>>> >>>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>>> This message (including any attachments) is for the named >>>> addressee(s)'s use only. It may contain sensitive, confidential, >>>> private proprietary or legally privileged information intended for a >>>> specific individual and purpose, and is protected by law. If you are >>>> not the intended recipient, please immediately delete it and all >>>> copies of it from your system, destroy any hard copies of it and >>>> notify the sender. Any use, disclosure, copying, or distribution of >>>> this message and/or any attachments is strictly prohibited. >>>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>>> >>>> -------------------------------------------------------------------- >>>> - >>>> - >>>> -------- _______________________________________________ >>>> Opensaf-users mailing list >>>> Ope...@li... >>>> https://lists.sourceforge.net/lists/listinfo/opensaf-users >>>> >>>> >>>> >>>> >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> This message (including any attachments) is for the named >>> addressee(s)'s use only. It may contain sensitive, confidential, >>> private proprietary or legally privileged information intended for a >>> specific individual and purpose, and is protected by law. If you are >>> not the intended recipient, please immediately delete it and all >>> copies of it from your system, destroy any hard copies of it and >>> notify the sender. Any use, disclosure, copying, or distribution of >>> this message and/or any attachments is strictly prohibited. >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >>> >>> >>> >>> >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> This message (including any attachments) is for the named >> addressee(s)'s use only. It may contain sensitive, confidential, >> private proprietary or legally privileged information intended for a >> specific individual and purpose, and is protected by law. If you are >> not the intended recipient, please immediately delete it and all >> copies of it from your system, destroy any hard copies of it and >> notify the sender. Any use, disclosure, copying, or distribution of >> this message and/or any attachments is strictly prohibited. >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >> >> >> > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all copies > of it from your system, destroy any hard copies of it > and notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Yao C. L. <yc...@as...> - 2014-07-15 11:42:08
|
Dear all, I noticed that there are AmfNodeName and ClmNodeName configured in imm.xml, as blow shows onf of my SC node, ------------------------------------------------------------------------------------------------------------------------- <object class="SaAmfNode"> <dn>safAmfNode=SC-2,safAmfCluster=myAmfCluster</dn> <attr> <name>saAmfNodeSuFailoverMax</name> <value>2</value> </attr> <attr> <name>saAmfNodeSuFailOverProb</name> <value>1200000000000</value> </attr> <attr> <name>saAmfNodeClmNode</name> <value>safNode=WR20-64_25,safCluster=myClmCluster</value> </attr> </object> ------------------------------------------------------------------------------------------------------------------------- root@WR20-64_32:/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>> cat node_name WR20-64_32 ------------------------------------------------------------------------------------------------------------------------- root@WR20-64_32:/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>> hostname WR20-64_32 ------------------------------------------------------------------------------------------------------------------------ I am curious to know when a node boots up, which it would read to get the clmNodeName, from :/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>/node_name or from hostname? And in readme_config, it says: ---------------------------------------------------------------------------------------------------------------------- 2) Edit nodes.cfg The third column in nodes.cfg should be edited to match the short form of the hostname (as shown by 'hostname -s') for each host. ----------------------------------------------------------------------------------------------------------------------- The what the second column and the third used for? Thanks. Ted ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Mathivanan N. P. <mat...@or...> - 2014-07-29 00:30:24
|
Hi, This had slipped through, Comments inline: >-----Original Message----- >From: Yao Cheng LIANG [mailto:yc...@as...] >Sent: Tuesday, July 15, 2014 5:12 PM >To: Anders Bjornerstedt >Cc: ope...@li...; santosh satapathy >Subject: [users] AmfNodeName ClmNodeName and hostname > >Dear all, > >I noticed that there are AmfNodeName and ClmNodeName configured in >imm.xml, as blow shows onf of my SC node, >----------------------------------------------------------------------------------------------- >-------------------------- > <object class="SaAmfNode"> > <dn>safAmfNode=SC-2,safAmfCluster=myAmfCluster</dn> > <attr> > <name>saAmfNodeSuFailoverMax</name> > <value>2</value> > </attr> > <attr> > <name>saAmfNodeSuFailOverProb</name> > <value>1200000000000</value> > </attr> > <attr> > <name>saAmfNodeClmNode</name> > <value>safNode=WR20-64_25,safCluster=myClmCluster</value> > </attr> > </object> >----------------------------------------------------------------------------------------------- >-------------------------- >root@WR20-64_32:/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>> >cat node_name >WR20-64_32 >----------------------------------------------------------------------------------------------- >-------------------------- >root@WR20-64_32:/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>> >hostname >WR20-64_32 >----------------------------------------------------------------------------------------------- >------------------------- > >I am curious to know when a node boots up, which it would read to get the >clmNodeName, from :/etc/opensaf<mailto:root@WR20- >64_32:/etc/opensaf>/node_name or from hostname? > By default, when you install opensaf, the default value in /etc/opensaf/node_name will be the output of the command 'hostname -s'. The value in /etc/opensaf/node_name represents the rdn of the CLM node name. >And in readme_config, it says: >----------------------------------------------------------------------------------------------- >----------------------- >2) Edit nodes.cfg > >The third column in nodes.cfg should be edited to match the short form of the >hostname (as shown by 'hostname -s') for each host. >----------------------------------------------------------------------------------------------- >------------------------ > >The what the second column and the third used for? > [Mathi] The first column is node_type, second column contains rdn value of AMF node name, and The 3rd column represents the rdn value of the CLM node name. Typically the CLM node name would be the hostname. The nodes.cfg is generated with default values set to SC-1, SC-2 etc. **This value should match The value that you specify in /etc/opensaf/node_name**. In your case, if you want safNode=WR20-64_25,safCluster=myClmCluster as your CLM node name, And if you want the AMF node name as safAmfNode= WR20-64_25,safAmfCluster=myAmfCluster, THEN You should configure /etc/opensaf/node_name with the value WR20-64_25 And your nodes.cfg should look like For eg:- SC WR20-64_25 WR20-64_25 Mathi. >Thanks. > >Ted > > >~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >This message (including any attachments) is for the named addressee(s)'s use >only. It may contain sensitive, confidential, private proprietary or legally >privileged information intended for a specific individual and purpose, and is >protected by law. If you are not the intended recipient, please immediately >delete it and all copies of it from your system, destroy any hard copies of it >and notify the sender. Any use, disclosure, copying, or distribution of this >message and/or any attachments is strictly prohibited. >~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > >------------------------------------------------------------------------------ >Want fast and easy access to all the code in your enterprise? Index and search >up to 200,000 lines of code with a free copy of Black Duck Code Sight - the >same software that powers the world's largest code search on Ohloh, the >Black Duck Open Hub! Try it now. >http://p.sf.net/sfu/bds >_______________________________________________ >Opensaf-users mailing list >Ope...@li... >https://lists.sourceforge.net/lists/listinfo/opensaf-users |
From: Yao C. L. <yc...@as...> - 2014-07-29 08:15:54
|
Thanks. It is very clearly explained. /Ted -----Original Message----- From: Mathivanan Naickan Palanivelu [mailto:mat...@or...] Sent: Tuesday, July 29, 2014 8:29 AM To: Yao Cheng LIANG; Anders Bjornerstedt Cc: ope...@li...; santosh satapathy Subject: RE: [users] AmfNodeName ClmNodeName and hostname Hi, This had slipped through, Comments inline: >-----Original Message----- >From: Yao Cheng LIANG [mailto:yc...@as...] >Sent: Tuesday, July 15, 2014 5:12 PM >To: Anders Bjornerstedt >Cc: ope...@li...; santosh satapathy >Subject: [users] AmfNodeName ClmNodeName and hostname > >Dear all, > >I noticed that there are AmfNodeName and ClmNodeName configured in >imm.xml, as blow shows onf of my SC node, >----------------------------------------------------------------------- >------------------------ >-------------------------- > <object class="SaAmfNode"> > <dn>safAmfNode=SC-2,safAmfCluster=myAmfCluster</dn> > <attr> > <name>saAmfNodeSuFailoverMax</name> > <value>2</value> > </attr> > <attr> > <name>saAmfNodeSuFailOverProb</name> > <value>1200000000000</value> > </attr> > <attr> > <name>saAmfNodeClmNode</name> > <value>safNode=WR20-64_25,safCluster=myClmCluster</value> > </attr> > </object> >----------------------------------------------------------------------- >------------------------ >-------------------------- >root@WR20-64_32:/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>> >cat node_name >WR20-64_32 >----------------------------------------------------------------------- >------------------------ >-------------------------- >root@WR20-64_32:/etc/opensaf<mailto:root@WR20-64_32:/etc/opensaf>> >hostname >WR20-64_32 >----------------------------------------------------------------------- >------------------------ >------------------------- > >I am curious to know when a node boots up, which it would read to get >the clmNodeName, from :/etc/opensaf<mailto:root@WR20- >64_32:/etc/opensaf>/node_name or from hostname? > By default, when you install opensaf, the default value in /etc/opensaf/node_name will be the output of the command 'hostname -s'. The value in /etc/opensaf/node_name represents the rdn of the CLM node name. >And in readme_config, it says: >----------------------------------------------------------------------- >------------------------ >----------------------- >2) Edit nodes.cfg > >The third column in nodes.cfg should be edited to match the short form >of the hostname (as shown by 'hostname -s') for each host. >----------------------------------------------------------------------- >------------------------ >------------------------ > >The what the second column and the third used for? > [Mathi] The first column is node_type, second column contains rdn value of AMF node name, and The 3rd column represents the rdn value of the CLM node name. Typically the CLM node name would be the hostname. The nodes.cfg is generated with default values set to SC-1, SC-2 etc. **This value should match The value that you specify in /etc/opensaf/node_name**. In your case, if you want safNode=WR20-64_25,safCluster=myClmCluster as your CLM node name, And if you want the AMF node name as safAmfNode= WR20-64_25,safAmfCluster=myAmfCluster, THEN You should configure /etc/opensaf/node_name with the value WR20-64_25 And your nodes.cfg should look like For eg:- SC WR20-64_25 WR20-64_25 Mathi. >Thanks. > >Ted > > >~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ >This message (including any attachments) is for the named >addressee(s)'s use only. It may contain sensitive, confidential, >private proprietary or legally privileged information intended for a >specific individual and purpose, and is protected by law. If you are >not the intended recipient, please immediately delete it and all copies >of it from your system, destroy any hard copies of it and notify the >sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. >~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > >----------------------------------------------------------------------- >------- Want fast and easy access to all the code in your enterprise? >Index and search up to 200,000 lines of code with a free copy of Black >Duck Code Sight - the same software that powers the world's largest >code search on Ohloh, the Black Duck Open Hub! Try it now. >http://p.sf.net/sfu/bds >_______________________________________________ >Opensaf-users mailing list >Ope...@li... >https://lists.sourceforge.net/lists/listinfo/opensaf-users ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Yao C. L. <yc...@as...> - 2014-07-31 12:28:48
Attachments:
imm.zip
|
Dear all, I have define a component named “PGW_pgw1” in SU “SC-1”, together with another two component “MME_mme1” and “SGW_sgw1”, script name are the same “app_script”. For for MME, the script is in “/root/astir/mme/mme1/opensaf”; for sgw, it is in “/root/astir/sgw/sgw1/opensaf”; for PGW, it is in “/root/astir/pgw/pgw1/opensaf”. Upon opensaf start, script for mme and sgw can be called successfully, but for PGW, it complains “ Script not readable or path wrong”. I am sure that the script is in “/root/astir/pgw/pgw1/opensaf” and executable. Could anyone help me to have a look at attached imm.xml and tell me where I swrong? Thanks. Ted ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Nagendra K. <nag...@or...> - 2014-07-31 12:56:45
|
Please change " safSmfBundlei=PGW_pgw1" to safSmfBundle=PGW_pgw1 at : <object class="SaAmfCompType"> <dn>safVersion=1,safCompType=PGW_pgw1</dn> <attr> <name>saAmfCtCompCategory</name> <value>1</value> </attr> <attr> <name>saAmfCtSwBundle</name> <value>safSmfBundlei=PGW_pgw1</value> </attr> > -----Original Message----- > From: Yao Cheng LIANG [mailto:yc...@as...] > Sent: 31 July 2014 17:59 > To: Mathivanan Naickan Palanivelu; Anders Bjornerstedt > Cc: ope...@li...; santosh satapathy > Subject: [users] Script not readable or path wrong > > Dear all, > > I have define a component named "PGW_pgw1" in SU "SC-1", together with > another two component "MME_mme1" and "SGW_sgw1", script name are the > same "app_script". For for MME, the script is in > "/root/astir/mme/mme1/opensaf"; for sgw, it is in > "/root/astir/sgw/sgw1/opensaf"; for PGW, it is in > "/root/astir/pgw/pgw1/opensaf". Upon opensaf start, script for mme and sgw > can be called successfully, but for PGW, it complains " Script not readable or > path wrong". I am sure that the script is in "/root/astir/pgw/pgw1/opensaf" and > executable. > > > Could anyone help me to have a look at attached imm.xml and tell me where I > swrong? > > Thanks. > > Ted > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named addressee(s)'s use > only. It may contain sensitive, confidential, private proprietary or legally > privileged information intended for a specific individual and purpose, and is > protected by law. If you are not the intended recipient, please immediately > delete it and all copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of this message > and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > |
From: Yao C. L. <yc...@as...> - 2014-07-31 13:35:51
|
Thanks. I just found the error, but took very long time. What tool are you relying on for this kind of errors? Thanks. Ted Sent from Windows Mail From: Nagendra Kumar Sent: Thursday, July 31, 2014 8:56 PM To: Yao Cheng LIANG, Mathivanan Naickan Palanivelu, Anders Bjornerstedt Cc: ope...@li..., santosh satapathy Please change " safSmfBundlei=PGW_pgw1" to safSmfBundle=PGW_pgw1 at : <object class="SaAmfCompType"> <dn>safVersion=1,safCompType=PGW_pgw1</dn> <attr> <name>saAmfCtCompCategory</name> <value>1</value> </attr> <attr> <name>saAmfCtSwBundle</name> <value>safSmfBundlei=PGW_pgw1</value> </attr> > -----Original Message----- > From: Yao Cheng LIANG [mailto:yc...@as...] > Sent: 31 July 2014 17:59 > To: Mathivanan Naickan Palanivelu; Anders Bjornerstedt > Cc: ope...@li...; santosh satapathy > Subject: [users] Script not readable or path wrong > > Dear all, > > I have define a component named "PGW_pgw1" in SU "SC-1", together with > another two component "MME_mme1" and "SGW_sgw1", script name are the > same "app_script". For for MME, the script is in > "/root/astir/mme/mme1/opensaf"; for sgw, it is in > "/root/astir/sgw/sgw1/opensaf"; for PGW, it is in > "/root/astir/pgw/pgw1/opensaf". Upon opensaf start, script for mme and sgw > can be called successfully, but for PGW, it complains " Script not readable or > path wrong". I am sure that the script is in "/root/astir/pgw/pgw1/opensaf" and > executable. > > > Could anyone help me to have a look at attached imm.xml and tell me where I > swrong? > > Thanks. > > Ted > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named addressee(s)'s use > only. It may contain sensitive, confidential, private proprietary or legally > privileged information intended for a specific individual and purpose, and is > protected by law. If you are not the intended recipient, please immediately > delete it and all copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of this message > and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |
From: Yao C. L. <yc...@as...> - 2014-07-14 08:35:19
|
Dear Anders, Does it the line "22:35:26 localhost osafimmd[11706]: Node 5020f request sync" show that SC-1 has received Sync request? By the way, how to relate node ID to node name? For example how can I know which know " Node 5020f" refers to? Thanks. Ted -----Original Message----- From: Anders Bjornerstedt [mailto:and...@er...] Sent: Monday, July 14, 2014 4:13 PM To: Yao Cheng LIANG Cc: ope...@li...; santosh satapathy Subject: Re: [users] One of the controller wait for sync Hi , The sync request from SC2 clearly reaches SC1. Is any sync started at SC1 ? I cant see because the syslog snippet from SC1 is minimal, truncated right after the request arrives. /Anders Bjornerstedt Yao Cheng LIANG wrote: > Dear all, > > I am using OpenSAF 4.2.2, and when I start SC-2 after SC-1, below message appears in /var/log/message on sc-1: > ------------------------------------------------------ > Jul 12 22:35:26 localhost osaffmd[11690]: Peer Node_id 328207 : EE_ID > safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 localhost osafimmd[11706]: New IMMND process is on > STANDBY Controller at 5020f Jul 12 22:35:26 localhost osafimmd[11706]: > IMMND on controller (not currently coord) requests sync Jul 12 > 22:35:26 localhost osafimmd[11706]: Node 5020f request sync > sync-pid:8930 epoch:0 > ---------------------------------------------------------------------- > -------------------------------------------------------- > > while on sc-2, below message appears in /var/log/message: > ---------------------------------------------------------------------- > -------------------------------------------------------- > Jul 12 22:35:26 WR20-64_32 opensafd: Starting OpenSAF Services Jul 12 > 22:35:26 WR20-64_32 osafdtmd[8860]: Started Jul 12 22:35:26 WR20-64_32 > /etc/redhat-lsb/lsb_start_daemon: osafdtmd startup - OK Jul 12 > 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 > 22:35:26 WR20-64_32 osafrded[8878]: Started > > Jul 12 22:35:26 WR20-64_32 osafrded[8878]: Started Jul 12 22:35:26 > WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafrded startup - OK Jul > 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 > 22:35:26 WR20-64_32 osafrded[8878]: rde@5030f<mailto:rde@5030f> has > active state => Standby role Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: > Started Jul 12 22:35:26 WR20-64_32 osaffmd[8897]: EE_ID : > safEE=Linux_os_hosting_clm_node,safHE=4500_slot_14,safDomain=domain_1 > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osaffmd > startup - OK Jul 12 22:35:26 WR20-64_32 > /etc/redhat-lsb/lsb_log_message: - OK Jul 12 22:35:26 WR20-64_32 > osafimmd[8913]: Started Jul 12 22:35:26 WR20-64_32 osafimmd[8913]: > Initialization Success, role STANDBY Jul 12 22:35:26 WR20-64_32 > /etc/redhat-lsb/lsb_start_daemon: osafimmd startup - OK Jul 12 > 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul 12 > 22:35:26 WR20-64_32 osafimmnd[8930]: Started Jul 12 22:35:26 > WR20-64_32 osafimmnd[8930]: Initialization Success Jul 12 22:35:26 > WR20-64_32 osafimmnd[8930]: Director Service is up Jul 12 22:35:26 > WR20-64_32 /etc/redhat-lsb/lsb_start_daemon: osafimmnd startup - OK > Jul 12 22:35:26 WR20-64_32 /etc/redhat-lsb/lsb_log_message: - OK Jul > 12 22:35:26 WR20-64_32 osafimmnd[8930]: SERVER STATE: > IMM_SERVER_ANONYMOUS --> IMM_SERVER_CLUSTER_WAITING Jul 12 22:35:26 > WR20-64_32 osafimmnd[8930]: SERVER STATE: IMM_SERVER_CLUSTER_WAITING > --> IMM_SERVER_LOADING_PENDING Jul 12 22:35:26 WR20-64_32 > osafimmnd[8930]: REQUESTING SYNC Jul 12 22:35:26 WR20-64_32 > osafimmnd[8930]: SERVER STATE: IMM_SERVER_LOADING_PENDING --> > IMM_SERVER_SYNC_PENDING Jul 12 22:35:26 WR20-64_32 osafimmnd[8930]: > NODE STATE-> IMM_NODE_ISOLATED Jul 12 22:35:46 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 20 > seconds Jul 12 22:36:06 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 40 seconds Jul 12 22:36:26 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 60 > seconds Jul 12 22:36:46 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 80 seconds Jul 12 22:37:06 WR20-64_32 > osafimmnd[8930]: REQUESTING SYNC AGAIN 1000 Jul 12 22:37:06 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 100 > seconds Jul 12 22:37:06 WR20-64_32 osafimmnd[8930]: Redundant sync > request, when IMM_NODE_ISOLATED Jul 12 22:37:16 WR20-64_32 > osafdtmd[8860]: DTM:dtm_comm_socket_recv() failed rc : 22 Jul 12 > 22:37:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 120 seconds Jul 12 22:37:46 WR20-64_32 osafimmnd[8930]: > This node still waiting to be sync'ed after 140 seconds Jul 12 > 22:37:52 WR20-64_32 osafimmd[8913]: IMMND DOWN on active controller f3 > detected at standby immd!! f2. Possible failover Jul 12 22:37:52 > WR20-64_32 osafimmd[8913]: Resend of fevs message 1855, will not mbcp > to peer IMMD Jul 12 22:37:52 WR20-64_32 osafimmd[8913]: Message > count:1856 + 1 != 1856 Jul 12 22:38:06 WR20-64_32 osafimmnd[8930]: > This node still waiting to be sync'ed after 160 seconds Jul 12 > 22:38:26 WR20-64_32 osafimmnd[8930]: This node still waiting to be > sync'ed after 180 seconds Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: > REQUESTING SYNC AGAIN 2000 Jul 12 22:38:46 WR20-64_32 osafimmnd[8930]: > This node still waiting to be sync'ed after 200 seconds Jul 12 > 22:38:46 WR20-64_32 osafimmnd[8930]: Redundant sync request, when > IMM_NODE_ISOLATED Jul 12 22:38:53 WR20-64_32 osafdtmd[8860]: DTM: add > New incoming connection to fd : 22 Jul 12 22:39:06 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 220 > seconds Jul 12 22:39:26 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 240 seconds Jul 12 22:39:46 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 260 > seconds Jul 12 22:40:06 WR20-64_32 osafimmnd[8930]: This node still > waiting to be sync'ed after 280 seconds Jul 12 22:40:26 WR20-64_32 > osafimmnd[8930]: REQUESTING SYNC AGAIN 3000 Jul 12 22:40:26 WR20-64_32 > osafimmnd[8930]: This node still waiting to be sync'ed after 300 > seconds Jul 12 22:40:26 WR20-64_32 osafimmnd[8930]: Redundant sync > request, when IMM_NODE_ISOLATED > ---------------------------------------------------------------------- > --------------------------------------------------------- > > But I reverse the order - i.e. to start sc-2 and then sc-2, both > controller can be started successfully > > Could anyone tell me what's wrong? > > Thanks. > > Ted > > > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > This message (including any attachments) is for the named > addressee(s)'s use only. It may contain sensitive, confidential, > private proprietary or legally privileged information intended for a > specific individual and purpose, and is protected by law. If you are > not the intended recipient, please immediately delete it and all > copies of it from your system, destroy any hard copies of it and > notify the sender. Any use, disclosure, copying, or distribution of > this message and/or any attachments is strictly prohibited. > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > ---------------------------------------------------------------------- > -------- _______________________________________________ > Opensaf-users mailing list > Ope...@li... > https://lists.sourceforge.net/lists/listinfo/opensaf-users > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ This message (including any attachments) is for the named addressee(s)'s use only. It may contain sensitive, confidential, private proprietary or legally privileged information intended for a specific individual and purpose, and is protected by law. If you are not the intended recipient, please immediately delete it and all copies of it from your system, destroy any hard copies of it and notify the sender. Any use, disclosure, copying, or distribution of this message and/or any attachments is strictly prohibited. ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ |