Hi Hoang,

On 2/10/2017 3:09 PM, Vo Minh Hoang wrote:
> If cpnd is temporary down only, we don't need clean up anything.
> If cpnd is permanently down, the bad effect of this proposal is that replica
> is not clean up. But if cpnd permanently down, we have to reboot node for
> recovering so I think this cleanup is not really necessary.
>
> I also checked this implementation with possible test cases and have not
> seen any side effect.
> Please consider it
We are observing new node_user_info  databases mismatch Errors, while 
testing multiple CPND restart
with this patch,I will do more debugging and update the root cause.

===========================================================================================================================

Apr 12 14:06:57 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f0500cf0, arg=0x7f86f0501ef0
*Apr 12 14:06:58 SC-1 osafckptd[27594]: ER 
cpd_proc_decrease_node_user_info failed - no user on node id 0x2020F*
Apr 12 14:06:58 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f0501750, arg=0x7f86f0501ef0
*Apr 12 14:06:59 SC-1 osafckptd[27594]: ER 
cpd_proc_decrease_node_user_info failed - no user on node id 0x2020F*
Apr 12 14:06:59 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f0503ab0, arg=0x7f86f0501ef0
Apr 12 14:07:00 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f0500c70, arg=0x7f86f0501ef0
Apr 12 14:07:01 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f0500930, arg=0x7f86f0501ef0
*Apr 12 14:07:03 SC-1 osafckptd[27594]: ER 
cpd_proc_decrease_node_user_info failed - no user on node id 0x2020*F
Apr 12 14:07:03 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f04fe3a0, arg=0x7f86f0501ef0
Apr 12 14:07:04 SC-1 osafckptd[27594]: NO cpnd_down_process:: Start 
CPND_RETENTION timer id = 0x7f86f0500cf0, arg=0x7f86f0501ef0

===========================================================================================================================

-AVM


On 4/12/2017 11:08 AM, A V Mahesh wrote:
> Hi Hoang,
>
> On 2/10/2017 3:09 PM, Vo Minh Hoang wrote:
>> Dear Mahesh,
>>
>> Based on what I saw, in this case, retention time cannot detect CPND
>> temporarily down because its pid changed.
> I will check that , I have some test cases based this retention time , 
> not sure how were they working.
>
> Can you please provide reproducible steps, I did look at ticket , but 
> looks complex ,
> if you have any application that reproduces the case please share.
>
> -AVM
>>
>> If cpnd is temporary down only, we don't need clean up anything.
>> If cpnd is permanently down, the bad effect of this proposal is that 
>> replica
>> is not clean up. But if cpnd permanently down, we have to reboot node 
>> for
>> recovering so I think this cleanup is not really necessary.
>>
>> I also checked this implementation with possible test cases and have not
>> seen any side effect.
>> Please consider it.
>>
>> Thank you and best regards,
>> Hoang
>>
>> -----Original Message-----
>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>> Sent: Friday, February 10, 2017 10:40 AM
>> To: Hoang Vo <hoang.m...@dektech.com.au>; zoran.milinko...@ericsson.com
>> Cc: opensaf-devel@lists.sourceforge.net
>> Subject: Re: [PATCH 1 of 1] cpd: to correct failover behavior of cpsv
>> [#1765] V5
>>
>> Hi Hoang,
>>
>> The CPD_CPND_DOWN_RETENTION  is to recognize, ether CPND temporarily 
>> down or
>> permanently down, this is started a CPND is down and based on
>> cpd_evt_proc_timer_expiry(), cpd recognize that the CPND is complete 
>> down
>> and do cleanup, else  cpnd rejoined with in 
>> CPD_CPND_DOWN_RETENTION_TIME ,
>> the CPD_CPND_DOWN_RETENTION is stoped.
>>
>> If we stop CPD_CPND_DOWN_RETENTION timer in cpd_process_cpnd_dow(), 
>> do cpd
>> recognize the CPD permanently down, the cpd_process_cpnd_dow() being 
>> called
>> in multiple flows, can you please check all the flows, is stopping
>> CPD_CPND_DOWN_RETENTION timer has any impact ?
>>
>> -AVM
>>
>> On 2/9/2017 1:35 PM, Hoang Vo wrote:
>>>    src/ckpt/ckptd/cpd_proc.c |  11 ++++++++++-
>>>    1 files changed, 10 insertions(+), 1 deletions(-)
>>>
>>>
>>> problem:
>>> In case failover multiple times, the cpnd is down for a moment so
>>> there is no cpnd opening specific checkpoint. This lead to retention 
>>> timer
>> is trigger.
>>> When cpnd is up again but has different pid so retention timer is not
>> stoped.
>>> Repica is deleted at retention while its information still be in ckpt
>> database.
>>> That cause problem
>>>
>>> Fix:
>>> - Stop timer of removed node.
>>> - Update data in patricia trees (for retention value consistence).
>>>
>>> diff --git a/src/ckpt/ckptd/cpd_proc.c b/src/ckpt/ckptd/cpd_proc.c
>>> --- a/src/ckpt/ckptd/cpd_proc.c
>>> +++ b/src/ckpt/ckptd/cpd_proc.c
>>> @@ -679,7 +679,8 @@ uint32_t cpd_process_cpnd_down(CPD_CB *c
>>>        cpd_cpnd_info_node_find_add(&cb->cpnd_tree, cpnd_dest, 
>>> &cpnd_info,
>> &add_flag);
>>>        if (!cpnd_info)
>>>            return NCSCC_RC_SUCCESS;
>>> -
>>> +    /* Stop timer before processing down */
>>> +    cpd_tmr_stop(&cpnd_info->cpnd_ret_timer);
>>>        cref_info = cpnd_info->ckpt_ref_list;
>>>           while (cref_info) {
>>> @@ -989,6 +990,14 @@ uint32_t cpd_proc_retention_set(CPD_CB *
>>>           /* Update the retention Time */
>>>        (*ckpt_node)->ret_time = reten_time;
>>> +    (*ckpt_node)->attributes.retentionDuration = reten_time;
>>> +
>>> +    /* Update the related patricia tree */
>>> +    CPD_CKPT_MAP_INFO *map_info = NULL;
>>> +    cpd_ckpt_map_node_get(&cb->ckpt_map_tree, (*ckpt_node)->ckpt_name,
>> &map_info);
>>> +    if (map_info) {
>>> +        map_info->attributes.retentionDuration = reten_time;
>>> +    }
>>>        return rc;
>>>    }
>>
>


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, Slashdot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to