Nikola Mandic created FLUME-2659:
------------------------------------
Summary: s3 sink unable to recover
Key: FLUME-2659
URL: https://issues.apache.org/jira/browse/FLUME-2659
Project: Flume
Issue Type: Bug
Components: Sinks+Sources
Affects Versions: v1.5.2
Environment: linux running on aws with hdfs library
https://github.com/twitter/hadoop-lzo.git
https://archive.apache.org/dist/hadoop/core/hadoop-2.4.0/hadoop-2.4.0.tar.gz
this is the docker file
````
RUN apt-get install -y wget default-jre git make curl vim tmux make gcc
autoconf build-essential libc-dev maven liblzo2-2 liblzo2-dev default-jdk
RUN cd ~
RUN wget
http://ftp.halifax.rwth-aachen.de/apache/flume/1.5.2/apache-flume-1.5.2-bin.tar.gz
RUN tar xf apache-flume-1.5.2-bin.tar.gz
ENV HOME /root
WORKDIR /root/apache-flume-1.5.2-bin
RUN rm -r /root/apache-flume-1.5.2-bin/conf
ADD ./conf /root/apache-flume-1.5.2-bin/conf
WORKDIR /root/
RUN git clone https://github.com/twitter/hadoop-lzo.git
RUN wget
https://archive.apache.org/dist/hadoop/core/hadoop-2.4.0/hadoop-2.4.0.tar.gz
RUN tar xf hadoop-2.4.0.tar.gz
WORKDIR /root/hadoop-2.4.0
RUN rm -r /root/hadoop-2.4.0/etc
ADD ./etc /root/hadoop-2.4.0/etc
WORKDIR /root/hadoop-lzo
ENV JAVA_HOME /usr/lib/jvm/java-7-openjdk-amd64
RUN C_INCLUDE_PATH=/usr/local/lzo-2.06/include
LIBRARY_PATH=/usr/local/lzo-2.06/lib mvn clean test
RUN mvn install
RUN cp target/hadoop-lzo-0.4.20-SNAPSHOT.jar ~/apache-flume-1.5.2-bin/lib/
ENV CLASSPATH
/usr/lib/jvm/java-7-openjdk-amd64/lib/:/root/apache-flume-1.5.2-bin/lib/*:/root/hadoop-2.4.0/share/hadoop/tools/lib/*:/root/.m2/repository/org/apache/hadoop/hadoop-common/2.4.0/hadoop-common-2.4.0.jar:/root/hadoop-2.4.0/etc/hadoop/
````
Reporter: Nikola Mandic
I use hdfs sink to write to s3.
Often s3 fails for some reason for example
```org.apache.flume.sink.hdfs.HDFSEventSink.process:463) - HDFS IO error```.
Sometimes it can be something else like reporting that it can not continue from
bad mark etc.
The problem is that hdfs sink does not recover unless I do manual restart.
What to do in those situations when s3 fails after running for long long time?
It would be quite nice if hdfs sink would restart connection completely instead
of me.
Right now it just bumps into error from time to time and stops working until
restart.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)