[ 
https://issues.apache.org/jira/browse/FLUME-1779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13553355#comment-13553355
 ] 

Connor Woodson commented on FLUME-1779:
---------------------------------------

The simple way to fix this that would enable the sink to work on the fail-over 
processor is to replace the return statement with throwing an exception. 
However, when the sink isn't used in a fail-over processor, backoff is a much 
more desirable return value, as there is a possibility that events may 
eventually be writable. There could be a config setting such as 
'retryBadConnection' defaulted to true that would run the return statement, 
thus keeping backwards functionality, but when false would throw the exception; 
is there a better way?
                
> Backoff returned from HDFSEventSink after IOException
> -----------------------------------------------------
>
>                 Key: FLUME-1779
>                 URL: https://issues.apache.org/jira/browse/FLUME-1779
>             Project: Flume
>          Issue Type: Bug
>            Reporter: Jaroslaw Grabowski
>         Attachments: st
>
>
> Status.Backoff is returned from HdfsEventSink.process() in case of 
> IOException. This behavior prevents FailoverSinkProcessor from pushing event 
> to next sink in queue.
> In my test case, IOException is caused by serious hdfs failure, for example 
> all DataNodes in cluster are dead. After such failure BucketWriter throws 
> IOException and becomes unavailable - it probably should be removed from 
> sfWriters map.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to