Hi!
Forgive the n00b question, but I'm trying to benchmark flume while building
out a hadoop based central log store and am coming across some weirdness.
The flume-conf.xml has the default flume.collector.output.format set to
'avrojson'. I had two simple configs:
test1: syslogTcp(5140) | collectorSink("hdfs://...", "test", 30000,
"avrodata")
test2: syslogTcp(5140) | collectorSink("hdfs://...", "test", 30000, "raw")
I then mapped a test flume node to each of these logical nodes in turn (exec
map node1 test1; exec refreshAll) and tested it out but the actual dfs files
are all appear to be the same size and all appear to be avronjson?
Am I doing something wrong here?
Using flume version: 0.9.4-cdh3u1.
Thanks,
-Steve
|