[ 
https://issues.apache.org/jira/browse/CHUKWA-4?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12852577#action_12852577
 ] 

Jerome Boulon commented on CHUKWA-4:
------------------------------------

You could create an helper class responsible for doing the recovery then you 
can call this helper class from the LocalToRemoteHDFSMover class.
But for the SeqFileWriter, it will be more difficult since your file will be 
visible by all collectors on HDFS so you would need to restrict the list of 
files per collector. So a better option for HDFS could be to have another 
deamon responsible for doing the recovery (wakeup every xx minutes and do a ls 
*.chukwa, check for last access time and recover and there's something to 
process).
 

> Collectors don't finish writing .done datasink from last .chukwa datasink 
> when stopped using bin/stop-collectors
> ----------------------------------------------------------------------------------------------------------------
>
>                 Key: CHUKWA-4
>                 URL: https://issues.apache.org/jira/browse/CHUKWA-4
>             Project: Hadoop Chukwa
>          Issue Type: Bug
>          Components: data collection
>         Environment: I am running on our local cluster. This is a linux 
> machine that I also run Hadoop cluster from.
>            Reporter: Andy Konwinski
>            Priority: Minor
>
> When I use start-collectors, it creates the datasink as expected, writes to 
> it as per normal, i.e. writes to the .chukwa file, and roll overs work fine 
> when it renames the .chukwa file to .done. However, when I use 
> bin/stop-collectors to shut down the running collector it leaves a .chukwa 
> file in the HDFS file system. Not sure if this is a valid sink or not, but I 
> think that the collector should gracefully clean up the datasink and rename 
> it .done before exiting.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to