Hi! Forgive the n00b question, but I'm trying to benchmark flume while building out a hadoop based central log store and am coming across some weirdness. The flume-conf.xml has the default flume.collector.output.format set to 'avrojson'. I had two simple configs:
test1: syslogTcp(5140) | collectorSink("hdfs://...", "test", 30000, "avrodata") test2: syslogTcp(5140) | collectorSink("hdfs://...", "test", 30000, "raw") I then mapped a test flume node to each of these logical nodes in turn (exec map node1 test1; exec refreshAll) and tested it out but the actual dfs files are all appear to be the same size and all appear to be avronjson? Am I doing something wrong here? Using flume version: 0.9.4-cdh3u1. Thanks, -Steve