Hi team I created flume source and sink as following in hadoop yarn and I am
not getting data transferred from source to sink in HDFS it doesnt create any
file and on local everytime I start agent it creates one empty file. Below are
my configs in source and sink
Source :-
agent.sources = logger1agent.sources.logger1.type =
execagent.sources.logger1.command = tail -f
/var/log/messagesagent.sources.logger1.batchsSize =
0agent.sources.logger1.channels = memoryChannelagent.channels =
memoryChannelagent.channels.memoryChannel.type =
memoryagent.channels.memoryChannel.capacity = 100agent.sinks =
AvroSinkagent.sinks.AvroSink.type = avroagent.sinks.AvroSink.channel =
memoryChannelagent.sinks.AvroSink.hostname =
192.168.147.101agent.sinks.AvroSink.port =
4545agent.sources.logger1.interceptors = itime
ihostagent.sources.logger1.interceptors.itime.type =
TimestampInterceptoragent.sources.logger1.interceptors.ihost.type =
hostagent.sources.logger1.interceptors.ihost.useIP =
falseagent.sources.logger1.interceptors.ihost.hostHeader = host
Sink at one of the slave ( datanodes on my Yarn cluster ) :
collector.sources = AvroIncollector.sources.AvroIn.type =
avrocollector.sources.AvroIn.bind = 0.0.0.0collector.sources.AvroIn.port =
4545collector.sources.AvroIn.channels = mc1 mc2collector.channels = mc1
mc2collector.channels.mc1.type = memorycollector.channels.mc1.capacity = 100
collector.channels.mc2.type = memorycollector.channels.mc2.capacity = 100
collector.sinks = LocalOut HadoopOutcollector.sinks.LocalOut.type =
file_rollcollector.sinks.LocalOut.sink.directory =
/home/hadoop/flumecollector.sinks.LocalOut.sink.rollInterval =
0collector.sinks.LocalOut.channel = mc1collector.sinks.HadoopOut.type =
hdfscollector.sinks.HadoopOut.channel = mc2collector.sinks.HadoopOut.hdfs.path
= /flumecollector.sinks.HadoopOut.hdfs.fileType =
DataStreamcollector.sinks.HadoopOut.hdfs.writeFormat =
Textcollector.sinks.HadoopOut.hdfs.rollSize =
0collector.sinks.HadoopOut.hdfs.rollCount =
10000collector.sinks.HadoopOut.hdfs.rollInterval = 600
can somebody point me to what am I doing wrong ?
This is what I get in my local directory
[hadoop@node1 flume]$ ls -lrttotal 0-rw-rw-r-- 1 hadoop hadoop 0 Oct 31 11:25
1383243942803-1-rw-rw-r-- 1 hadoop hadoop 0 Oct 31 11:28
1383244097923-1-rw-rw-r-- 1 hadoop hadoop 0 Oct 31 11:31
1383244302225-1-rw-rw-r-- 1 hadoop hadoop 0 Oct 31 11:33 1383244404929-1
when I restart the collector it creates one 0 bytes file.
Please help
*------------------------*
Cheers !!!
Siddharth Tiwari
Have a refreshing day !!!
"Every duty is holy, and devotion to duty is the highest form of worship of
God.”
"Maybe other people will try to limit me but I don't limit myself"