Hi!

Forgive the n00b question, but I'm trying to benchmark flume while building
out a hadoop based central log store and am coming across some weirdness.
 The flume-conf.xml has the default flume.collector.output.format set to
'avrojson'.  I had two simple configs:

test1: syslogTcp(5140) | collectorSink("hdfs://...", "test", 30000,
"avrodata")
test2: syslogTcp(5140) | collectorSink("hdfs://...", "test", 30000, "raw")

I then mapped a test flume node to each of these logical nodes in turn (exec
map node1 test1; exec refreshAll) and tested it out but the actual dfs files
are all appear to be the same size and all appear to be avronjson?

Am I doing something wrong here?

Using flume version: 0.9.4-cdh3u1.

Thanks,

-Steve

Reply via email to