[ 
https://issues.apache.org/jira/browse/HADOOP-5730?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12702198#action_12702198
 ] 

Wang Xu commented on HADOOP-5730:
---------------------------------

>   1. What happens if all directories are removed on SecondareNameNode?

It's quite a problem, and do you think SecondaryNameNode should throw Exception 
or kill itself ?

>   2. Why do you remove directories only if mkdir() fails? What if rename() 
> fails before mkdir() for example.

I think rename maybe should also be "try...catch"

>   3. You cannot just remove a list entry while iterating, this will cause 
> ConcurrentModificationException 
>   on the next iteration of the loop.

oh. I am sorry for that, I will change its position.

I will modify it and upload another patch. And I wonder whether it is OK  if we 
only record this problem
in logfiles and ignore it.

> SecondaryNameNode:  should not throw exception and exit if only one makedir 
> failure
> -----------------------------------------------------------------------------------
>
>                 Key: HADOOP-5730
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5730
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.19.1
>            Reporter: Wang Xu
>            Assignee: Wang Xu
>             Fix For: 0.19.2
>
>         Attachments: secondarynamenode-startcp.patch
>
>   Original Estimate: 2h
>  Remaining Estimate: 2h
>
> In CheckpointStorage.startCheckPointing(), if one mkdir failed, it 
> will throw an exception and exit. 
> However, because the editlog has been closed before, the editStreams
> of FSEditLog of NameNode will becomes empty as a result, which
> will affect any further logSync operations.
> Hence we think it should only print  WARN message instead of 
> throw the exception

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to