Hi Minh,

 

Please see my responses inline.

In general, I am trying to implement new version to reuse exist database.

 

Best Regards,

ThuanTr

 

From: Minh Hon Chau <minh.c...@dektech.com.au> 
Sent: Wednesday, October 23, 2019 12:23 PM
To: Tran Thuan <thuan.t...@dektech.com.au>; 'Nguyen Minh Vu' 
<vu.m.ngu...@dektech.com.au>; hans.nordeb...@ericsson.com; 
gary....@dektech.com.au
Cc: opensaf-devel@lists.sourceforge.net
Subject: Re: [PATCH 1/1] mds: not waste 1.5s in waiting Adest already down to 
send response message [#3102]

 

Hi Thuan,

Please see comment inline

Thanks

Minh

On 23/10/19 3:32 pm, Tran Thuan wrote:

Hi Minh,

 

Thanks for comments. See my response inline.

Btw, I am preparing to send out new patch, I think I found an issue in current 
patch.

 

Best Regards,

ThuanTr

 

-----Original Message-----
From: Minh Hon Chau  <mailto:minh.c...@dektech.com.au> 
<minh.c...@dektech.com.au> 
Sent: Wednesday, October 23, 2019 5:52 AM
To: Tran Thuan  <mailto:thuan.t...@dektech.com.au> <thuan.t...@dektech.com.au>; 
'Nguyen Minh Vu'  <mailto:vu.m.ngu...@dektech.com.au> 
<vu.m.ngu...@dektech.com.au>; hans.nordeb...@ericsson.com 
<mailto:hans.nordeb...@ericsson.com> ; gary....@dektech.com.au 
<mailto:gary....@dektech.com.au> 
Cc: opensaf-devel@lists.sourceforge.net 
<mailto:opensaf-devel@lists.sourceforge.net> 
Subject: Re: [PATCH 1/1] mds: not waste 1.5s in waiting Adest already down to 
send response message [#3102]

 

Hi Thuan,

 

I wonder the patch would work in the same reproduced steps if the both 

adests have subscribed each other more than 2 services. The svc_cnt will 

be greater than 1 until it is the last service down event. I think 

that's why mds has the database @subtn_results, in which each item is an 

adest associated with a service id separately.

[Thuan] We can understand that adest still alive, then go with origin flow 
(wait 1.5s).

But can a process send SNDRSP then mds unregister? I think it cannot, because 
it’s in SNDRSP (blocking)

[M] Not unregister, it can be unsubscribe. Or do you mean a process can not 
send two SNDRSP at the same time on 2 different subscribed services? 
[T] I think process cannot do anything relate to MDS API because it is under 
send SNDRSP (blocking).

The scenario of this ticket happens for the process terminated/crash.

[M] Yes my doubt is in the context of this ticket - terminated/crash - you 
would get 2 service down event I think

[T] Yes, will get 2 service down events.

[M] I don't think adding a new database at the global scope for this specific 
case is a good idea, if we can reuse the existing database. Can you try to use 
MDS_SUBSCRIPTION_INFO, add a flag or something similar to indicate which case 
mds should wait 1.5 sec. It would isolate the bug fix in the scope of this 
problem.

[T] I don’t want to mess up current database and its logic. But anyway, I am 
trying to reuse that database. Please wait for next version.

 

The problem originally resides at the services code e.g ntf, imm... 

where the threads structure between mds receiving thread and main thread 

cause a race condition. Thus the service sends a message with a death 

adest which is removed from mds database, that confuses mds and hit 1.5 

secs wait time.

 

If I read the code correctly, the 1.5 wait time is for another case, it 

gives another chance to wait 1.5 when the subscription result is empty 

in @subtn_results because the service up has not arrived yet.

[Thuan] Yes, it will give a chance if adest not yet UP any.

My patch still keep that chance as origin code.

But I think I need reduce timeout for adest down timer, I am verifying this 
change.

 

mds ---- subscribe ---->

 

mds ---- sends message A ----x

 

mds wait 1.5 sec

 

mds <--- service up ----

 

mds ---- sends message A ---->

 

So the 1.5 sec time is for early phase of waiting service up.

 

    } else if (sub_info->tmr_flag != true) {

        if ((MDS_SENDTYPE_RSP == req->i_sendtype) ||

            (MDS_SENDTYPE_RRSP == req->i_sendtype)) {

            time_wait = true;

            m_MDS_LOG_INFO(

                "MDS_SND_RCV:Disc queue: Subscr exists no timer 

running: Waiting for some time\n");

 

-> I think at this line, it means: the SUBSCRIPTION_TMR has timeout, and 

mds is sending RSP/RRSP which means mds should have received the 

*request* message (?), so mds wants to wait for another 1.5 second for 

service_up to create the subscription result in database.

 

The problem in this ticket hit 1.5 because the service down has already 

come and mds removed the subscription result item, now the ntf, imm... 

sends msg with a death adest, and mds now it thinks it is waiting for a 

service up to come as at the early phase, so it waits. Both two 

scenarios can be distinguished themselves to avoid to wait 1.5 secs for 

the latter case I think.

 

Thanks

 

Minh

 

On 22/10/19 9:50 pm, Tran Thuan wrote:

> Hi Vu,

> 

> Thanks for additional comments.

> I reply your concerns inline below.

> 

> Best Regards,

> ThuanTr

> 

> -----Original Message-----

> From: Nguyen Minh Vu < <mailto:vu.m.ngu...@dektech.com.au> 
> vu.m.ngu...@dektech.com.au>

> Sent: Tuesday, October 22, 2019 5:28 PM

> To: thuan.tran < <mailto:thuan.t...@dektech.com.au> 
> thuan.t...@dektech.com.au>; 'Minh Chau' < <mailto:minh.c...@dektech.com.au> 
> minh.c...@dektech.com.au>;  <mailto:hans.nordeb...@ericsson.com> 
> hans.nordeb...@ericsson.com;  <mailto:gary....@dektech.com.au> 
> gary....@dektech.com.au

> Cc:  <mailto:opensaf-devel@lists.sourceforge.net> 
> opensaf-devel@lists.sourceforge.net

> Subject: Re: [PATCH 1/1] mds: not waste 1.5s in waiting Adest already down to 
> send response message [#3102]

> 

> Hi Thuan,

> 

> I have additional comments below.

> 

> Regards, Vu

> 

> On 10/22/19 7:14 AM, thuan.tran wrote:

>> - When sending response message which Adest not exist (already down)

>> current MDS try to wait for 1.5 seconds before conclude no route to

>> send response message.

>> 

>> - There are 2 scenarios may have:

>> UP -> DOWN -> receive SNDRSP -> response timeout after 1.5s

>> UP -> receive SNDRSP -> DOWN -> response timeout after 1.5s

>> 

>> - With this change, MDS will not waste for 1.5s which can cause trouble

>> for higher layer services, e.g: ntf, imm, etc...

>> ---

>>    src/mds/mds_c_api.c     | 70 ++++++++++++++++++++++++++++++++++++++++-

>>    src/mds/mds_c_sndrcv.c  | 52 ++++++++++++++++++++++++++++--

>>    src/mds/mds_core.h      | 25 +++++++++++++--

>>    src/mds/mds_dt2c.h      |  2 +-

>>    src/mds/mds_dt_common.c | 22 ++++++++++++-

>>    5 files changed, 162 insertions(+), 9 deletions(-)

>> 

>> diff --git a/src/mds/mds_c_api.c b/src/mds/mds_c_api.c

>> index 132555b8e..5dd30c536 100644

>> --- a/src/mds/mds_c_api.c

>> +++ b/src/mds/mds_c_api.c

>> @@ -1900,6 +1900,32 @@ uint32_t mds_mcm_svc_up(PW_ENV_ID pwe_id, MDS_SVC_ID 
>> svc_id, V_DEST_RL role,

>>    

>>           /*************** Validation for SCOPE **********************/

>>    

>> +       if (adest != m_MDS_GET_ADEST) {

>> +                       MDS_ADEST_INFO *adest_info =

>> +                                       (MDS_ADEST_INFO 
>> *)ncs_patricia_tree_get(

>> +                                                       
>> &gl_mds_mcm_cb->adest_list,

>> +                                                       (uint8_t *)&adest);

>> +                       if (!adest_info) {

>> +                                       /* Add adest to adest list */

>> +                                       adest_info = m_MMGR_ALLOC_ADEST_INFO;

>> +                                       memset(adest_info, 0, 
>> sizeof(MDS_ADEST_INFO));

>> +                                       adest_info->adest = adest;

>> +                                       adest_info->node.key_info =

>> +                                                       (uint8_t 
>> *)&adest_info->adest;

>> +                                       adest_info->svc_cnt = 1;

>> +                                       adest_info->tmr_start = false;

>> +                                       ncs_patricia_tree_add(

>> +                                           &gl_mds_mcm_cb->adest_list,

>> +                                           (NCS_PATRICIA_NODE *)adest_info);

>> +                                       m_MDS_LOG_DBG("MCM:API: Adest=%" 
>> PRIu64

>> +                                                       " svc_cnt=%u", 
>> adest, adest_info->svc_cnt);

>> +                       } else {

>> +                                       adest_info->svc_cnt++;

>> +                                       m_MDS_LOG_DBG("MCM:API: Adest=%" 
>> PRIu64

>> +                                                       " svc_cnt=%u", 
>> adest, adest_info->svc_cnt);

>> +                       }

>> +       }

>> +

>>           status = mds_get_subtn_res_tbl_by_adest(local_svc_hdl, svc_id, 
>> vdest_id,

>>                                                                              
>>              adest, &log_subtn_result_info);

>>    

>> @@ -3571,6 +3597,24 @@ uint32_t mds_mcm_svc_down(PW_ENV_ID pwe_id, 
>> MDS_SVC_ID svc_id, V_DEST_RL role,

>>                           /* Discard : Getting down before getting up */

>>           } else { /* Entry exist in subscription result table */

>>    

>> +                       /* If adest exist and no sndrsp, start a timer */

>> +                       MDS_ADEST_INFO *adest_info =

>> +                                       (MDS_ADEST_INFO 
>> *)ncs_patricia_tree_get(

>> +                                                       
>> &gl_mds_mcm_cb->adest_list,

>> +                                                       (uint8_t *)&adest);

>> +                       if (adest_info) {

>> +                                       adest_info->svc_cnt--;

>> +                                       if (adest_info->svc_cnt == 0 &&

>> +                                           adest_info->sndrsp_cnt == 0) {

>> +                                                       
>> m_MDS_LOG_INFO("MCM:API: Adest=%" PRIu64

>> +                                                                       " 
>> down timer start", adest);

>> +                                                       if 
>> (adest_info->tmr_start == false) {

>> +                                                                       
>> adest_info->tmr_start = true;

>> +                                                                       
>> start_mds_down_tmr(adest, svc_id);

>> +                                                       }

>> +                                       }

>> +                       }

>> +

>>                           if (vdest_id == m_VDEST_ID_FOR_ADEST_ENTRY) {

>>                                           status = mds_subtn_res_tbl_del(

>>                                               local_svc_hdl, svc_id, 
>> vdest_id, adest,

>> @@ -4956,6 +5000,17 @@ uint32_t mds_mcm_init(void)

>>                           return NCSCC_RC_FAILURE;

>>           }

>>    

>> +       /* ADEST TREE */

>> +       memset(&pat_tree_params, 0, sizeof(NCS_PATRICIA_PARAMS));

>> +       pat_tree_params.key_size = sizeof(MDS_DEST);

>> +       if (NCSCC_RC_SUCCESS !=

>> +           ncs_patricia_tree_init(&gl_mds_mcm_cb->adest_list,

>> +                                                          
>> &pat_tree_params)) {

>> +                       m_MDS_LOG_ERR(

>> +                           "MCM:API: patricia_tree_init: adest :failure, L 
>> mds_mcm_init");

>> +                       return NCSCC_RC_FAILURE;

>> +       }

>> +

>>           /* SERVICE TREE */

>>           memset(&pat_tree_params, 0, sizeof(NCS_PATRICIA_PARAMS));

>>           pat_tree_params.key_size = sizeof(MDS_SVC_HDL);

>> @@ -4966,7 +5021,12 @@ uint32_t mds_mcm_init(void)

>>                           if (NCSCC_RC_SUCCESS !=

>>                               
>> ncs_patricia_tree_destroy(&gl_mds_mcm_cb->vdest_list)) {

>>                                           m_MDS_LOG_ERR(

>> -                                            "MCM:API: 
>> patricia_tree_destroy: service :failure, L mds_mcm_init");

>> +                                           "MCM:API: patricia_tree_destroy: 
>> vdest :failure, L mds_mcm_init");

>> +                       }

>> +                       if (NCSCC_RC_SUCCESS !=

>> +                           
>> ncs_patricia_tree_destroy(&gl_mds_mcm_cb->adest_list)) {

>> +                                       m_MDS_LOG_ERR(

>> +                                           "MCM:API: patricia_tree_destroy: 
>> adest :failure, L mds_mcm_init");

>>                           }

>>                           return NCSCC_RC_FAILURE;

>>           }

>> @@ -4989,6 +5049,11 @@ uint32_t mds_mcm_init(void)

>>                                           m_MDS_LOG_ERR(

>>                                               "MCM:API: 
>> patricia_tree_destroy: vdest :failure, L mds_mcm_init");

>>                           }

>> +                       if (NCSCC_RC_SUCCESS !=

>> +                           
>> ncs_patricia_tree_destroy(&gl_mds_mcm_cb->adest_list)) {

>> +                                       m_MDS_LOG_ERR(

>> +                                           "MCM:API: patricia_tree_destroy: 
>> adest :failure, L mds_mcm_init");

>> +                       }

>>                           return NCSCC_RC_FAILURE;

>>           }

>>    

>> @@ -5036,6 +5101,9 @@ uint32_t mds_mcm_destroy(void)

>>           /* VDEST TREE */

>>           ncs_patricia_tree_destroy(&gl_mds_mcm_cb->vdest_list);

>>    

>> +       /* ADEST TREE */

>> +       ncs_patricia_tree_destroy(&gl_mds_mcm_cb->adest_list);

>> +

>>           /* Free MCM control block */

>>           m_MMGR_FREE_MCM_CB(gl_mds_mcm_cb);

>>    

>> diff --git a/src/mds/mds_c_sndrcv.c b/src/mds/mds_c_sndrcv.c

>> index 7850ac714..4d8865b3a 100644

>> --- a/src/mds/mds_c_sndrcv.c

>> +++ b/src/mds/mds_c_sndrcv.c

>> @@ -2650,6 +2650,22 @@ static uint32_t mcm_pvt_red_snd_process_common(

>>                           }

>>           }

>>    

>> +       /* Update adest sndrsp counter */

>> +       if ((MDS_SENDTYPE_RSP == req->i_sendtype) ||

>> +           (MDS_SENDTYPE_RRSP == req->i_sendtype)) {

>> +                       MDS_ADEST_INFO *adest_info =

>> +                                       (MDS_ADEST_INFO 
>> *)ncs_patricia_tree_get(

>> +                                                       
>> &gl_mds_mcm_cb->adest_list, (uint8_t *)&dest);

>> +                       if (adest_info) {

>> +                                       adest_info->sndrsp_cnt--;

> [Vu] since this routine is triggered by mds's user, so if user makes a

> mistake that calling MDS_SENDTYPE_RSP | MDS_SENDTYPE_RRSP

> in case mds not yet receives MDS_SENDTYPE_SNDRSP | MDS_SENDTYPE_REDRSP.

> If that is the case, sndrsp_cnt counter could reach the max uint32 (-1).

> [Thuan] To send RSP user must in a SNDRSP context, invalid call will be 
> detected before come here.

>> +                                       if (adest_info->sndrsp_cnt == 0 &&

>> +                                           adest_info->svc_cnt == 0)

>> +                                                       
>> ncs_patricia_tree_del(

>> +                                                                       
>> &gl_mds_mcm_cb->adest_list,

>> +                                                                       
>> (NCS_PATRICIA_NODE *)adest_info);

>> +                       }

>> +       }

>> +

>>           if (NCSCC_RC_SUCCESS != mds_subtn_res_tbl_get_by_adest(

>>                                                               
>> svc_cb->svc_hdl, to_svc_id, dest_vdest_id,

>>                                                               dest, 
>> &role_ret, &subs_result_hdl)) {

>> @@ -2716,9 +2732,27 @@ static uint32_t 
>> mds_mcm_process_disc_queue_checks_redundant(

>>           } else if (sub_info->tmr_flag != true) {

>>                           if ((MDS_SENDTYPE_RSP == req->i_sendtype) ||

>>                               (MDS_SENDTYPE_RRSP == req->i_sendtype)) {

>> -                                        time_wait = true;

>> -                                        m_MDS_LOG_INFO(

>> -                                            "MDS_SND_RCV:Disc queue red: 
>> Subscr exists no timer running: Waiting for some time\n");

>> +                                       MDS_ADEST_INFO *adest_info =

>> +                                                       (MDS_ADEST_INFO 
>> *)ncs_patricia_tree_get(

>> +                                                                       
>> &gl_mds_mcm_cb->adest_list,

>> +                                                                       
>> (uint8_t *)&anchor);

>> +                                       if (adest_info && 
>> adest_info->svc_cnt == 0) {

>> +                                                       
>> adest_info->sndrsp_cnt--;

>> +                                                       m_MDS_LOG_NOTIFY(

>> +                                                                       
>> "MDS_SND_RCV: Adest already down,"

>> +                                                                       " 
>> skip response, remain sndrsp_cnt=%d",

>> +                                                                       
>> adest_info->sndrsp_cnt);

>> +                                                       if 
>> (adest_info->sndrsp_cnt == 0)

>> +                                                                       
>> ncs_patricia_tree_del(

>> +                                                                           
>> &gl_mds_mcm_cb->adest_list,

>> +                                                                           
>> (NCS_PATRICIA_NODE *)adest_info);

>> +                                                       return 
>> NCSCC_RC_FAILURE;

> [Vu] What if adest_info is null? we should return ok to caller because

> that in that case adest is down.

> [Thuan] adest_info is null, below else will run. (UP may not yet come)

>> +                                       } else {

>> +                                                       time_wait = true;

>> +                                                       m_MDS_LOG_INFO(

>> +                                                           
>> "MDS_SND_RCV:Disc queue red: Subscr exists"

>> +                                                           " no timer 
>> running: Waiting for some time");

>> +                                       }

>>                           } else {

>>                                           m_MDS_LOG_INFO(

>>                                               "MDS_SND_RCV: Subscription 
>> exists but Timer has expired\n");

>> @@ -5024,6 +5058,18 @@ static uint32_t 
>> mds_mcm_process_recv_snd_msg_common(MDS_SVC_INFO *svccb,

>>                                           mds_mcm_free_msg_memory(recv->msg);

>>                                           return NCSCC_RC_FAILURE;

>>                           }

>> +

>> +                       /* Update adest sndrsp counter */

>> +                       if (recv->snd_type == MDS_SENDTYPE_SNDRSP ||

>> +                           recv->snd_type == MDS_SENDTYPE_REDRSP) {

>> +                                       MDS_ADEST_INFO *adest_info =

>> +                                                       (MDS_ADEST_INFO 
>> *)ncs_patricia_tree_get(

>> +                                                                       
>> &gl_mds_mcm_cb->adest_list,

>> +                                                                       
>> (uint8_t *)&recv->src_adest);

>> +                                       if (adest_info)

>> +                                                       
>> adest_info->sndrsp_cnt++;

>> +                       }

>> +

>>                           cbinfo.info.receive.pid = recv->pid;

>>                           cbinfo.info.receive.uid = recv->uid;

>>                           cbinfo.info.receive.gid = recv->gid;

>> diff --git a/src/mds/mds_core.h b/src/mds/mds_core.h

>> index c09b428ba..f5d4d22a1 100644

>> --- a/src/mds/mds_core.h

>> +++ b/src/mds/mds_core.h

>> @@ -204,12 +204,12 @@ typedef struct mds_subscription_info {

>>    #define m_GET_HDL_FROM_MDS_SVC_INFO(info) \

>>      ((info->svc-id)<<32 ||                        \

>>       (info->parent_pwe->pwe_id)<<16 ||            \

>> -   (info->parent_pwe->parent_vdest->vdest-id)))

>> +   (info->parent_pwe->parent_vdest->vdest-id))

>>    

>>    #define m_GET_HDL_FROM_MDS_SVC_PWE_VDEST(s, p, v) \

>>      ((s)<<32 ||                                   \

>>       (p)<<16 ||                                   \

>> -   (v)))

>> +   (v))

>>    

>>    /**************************************\

>>        MDS PWE related declarations

>> @@ -251,6 +251,16 @@ typedef struct mds_vdest_info {

>>    

>>    } MDS_VDEST_INFO;

>>    

>> +typedef struct mds_adest_info {

>> +  /* Indexing info */

>> +  NCS_PATRICIA_NODE node;

>> +  /* Adest info */

>> +  MDS_DEST adest; /* Key for Patricia node */

>> +  uint16_t svc_cnt; /* Adest SVC counter */

>> +  uint32_t sndrsp_cnt; /* Counter for adest send sndrsp */

>> +  bool tmr_start; /* Adest down timer start flag */

>> +} MDS_ADEST_INFO;

>> +

>>    typedef struct mds_svc_info {

>>      /* Indexing info */

>>      NCS_PATRICIA_NODE svc_list_node;

>> @@ -301,6 +311,7 @@ typedef struct mds_mcm_cb {

>>      NCS_PATRICIA_TREE subtn_results;

>>      NCS_PATRICIA_TREE svc_list;   /* Tree of MDS_SVC_INFO information */

>>      NCS_PATRICIA_TREE vdest_list; /* Tree of MDS_VDEST_INFO information */

>> +  NCS_PATRICIA_TREE adest_list; /* Tree of MDS_ADEST_INFO information */

>>    } MDS_MCM_CB;

>>    

>>    /* Global MDSCB */

>> @@ -681,6 +692,15 @@ uint32_t (*mds_mdtm_node_unsubscribe)(MDS_SUBTN_REF_VAL 
>> subtn_ref_val);

>>      m_NCS_MEM_FREE(p, NCS_MEM_REGION_TRANSIENT, NCS_SERVICE_ID_MDS, \

>>                     MDS_MEM_VDEST_INFO)

>>    

>> +#define m_MMGR_ALLOC_ADEST_INFO                               \

>> +  (MDS_ADEST_INFO *)m_NCS_MEM_ALLOC(sizeof(MDS_ADEST_INFO),   \

>> +                                    NCS_MEM_REGION_TRANSIENT, \

>> +                                    NCS_SERVICE_ID_MDS, MDS_MEM_ADEST_INFO)

>> +

>> +#define m_MMGR_FREE_ADEST_INFO(p)                                 \

>> +  m_NCS_MEM_FREE(p, NCS_MEM_REGION_TRANSIENT, NCS_SERVICE_ID_MDS, \

>> +                 MDS_MEM_ADEST_INFO)

>> +

>>    #define m_MMGR_ALLOC_PWE_INFO                               \

>>      (MDS_PWE_INFO *)m_NCS_MEM_ALLOC(sizeof(MDS_PWE_INFO),     \

>>                                      NCS_MEM_REGION_TRANSIENT, \

>> @@ -768,7 +788,6 @@ typedef struct mds_mcm_msg_elem {

>>          NCSMDS_CALLBACK_INFO cbinfo;

>>        } event;

>>      } info;

>> -

>>    } MDS_MCM_MSG_ELEM;

>>    

>>    /* ******************************************** */

>> diff --git a/src/mds/mds_dt2c.h b/src/mds/mds_dt2c.h

>> index c92fecbba..49c9b09c4 100644

>> --- a/src/mds/mds_dt2c.h

>> +++ b/src/mds/mds_dt2c.h

>> @@ -223,7 +223,7 @@ typedef enum {

>>      MDS_MEM_DIRECT_BUFF,

>>      MDS_MEM_AWAIT_ACTIVE,

>>      MDS_MEM_MSGELEM,

>> -  MDS_MEM_ADEST_LIST,

>> +  MDS_MEM_ADEST_INFO,

>>      MDS_MEM_REASSEM_QUEUE,

>>      MDS_MEM_HDL_LIST,

>>      MDS_MEM_CACHED_EVENTS_LIST,

>> diff --git a/src/mds/mds_dt_common.c b/src/mds/mds_dt_common.c

>> index de1388367..19000b387 100644

>> --- a/src/mds/mds_dt_common.c

>> +++ b/src/mds/mds_dt_common.c

>> @@ -998,6 +998,26 @@ uint32_t mds_tmr_mailbox_processing(void)

>>                                                                              
>>  info->mds_dest, info->pid);

>>                                                                           
>> (void)mds_process_info_del(info);

>>                                                                           
>> free(info);

>> +                                                       } else {

>> +                                                                       
>> MDS_DEST adest =

>> +                                                                            
>>            tmr_req_info->info

>> +                                                                            
>>            .down_event_tmr_info.adest;

>> +                                                                       
>> m_MDS_LOG_INFO(

>> +                                                                           
>> "TIMEOUT Adest=%" PRIu64

>> +                                                                           
>> " down", adest);

>> +                                                                       
>> MDS_ADEST_INFO *adest_info =

>> +                                                                           
>> (MDS_ADEST_INFO *)

>> +                                                                           
>> ncs_patricia_tree_get(

>> +                                                                           
>> &gl_mds_mcm_cb->adest_list,

>> +                                                                           
>> (uint8_t *)&adest);

>> +                                                                       if 
>> (adest_info) {

>> +                                                                            
>>            adest_info->tmr_start = false;

>> +                                                                            
>>            if (adest_info->svc_cnt == 0 &&

>> +                                                                            
>>                adest_info->sndrsp_cnt == 0)

>> +                                                                            
>>                ncs_patricia_tree_del(

>> +                                                                            
>>                    &gl_mds_mcm_cb->adest_list,

>> +                                                                            
>>                    (NCS_PATRICIA_NODE *)adest_info);

>> +                                                                       }

>>                                                           }

>>    

>>                                                           if 
>> (tmr_req_info->info.down_event_tmr_info

>> @@ -1700,7 +1720,7 @@ uint16_t mds_checksum(uint32_t length, uint8_t buff[])

>>           /* Take the one's complement of sum */

>>           sum = ~sum;

>>    

>> -        return ((uint16_t)sum);

>> +       return (uint16_t)sum;

>>    }

>>    

>>    
>> /****************************************************************************

> 

> 


_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to