Hi, I have just compiled kafka from https://github.com/kafka-dev/kafka and executed the DataGenerator:
./run-class.sh kafka.etl.impl.DataGenerator test/test.properties After that I have executed the hadoop consumer: ./run-class.sh kafka.etl.impl.SimpleKafkaETLJob test/test.properties The hadoop-consumer is generating a file on the specified output but it is never finishing, even if I try to generate only 1 event at test/test.properties. So this file is growing and growing, my guessing is that maybe it is reading always the offset 0? That is my test.properties: # name of test topic kafka.etl.topic=SimpleTestEvent5 # hdfs location of jars hdfs.default.classpath.dir=/tmp/kafka/lib # number of test events to be generated event.count=1 # hadoop id and group hadoop.job.ugi=kafka,hadoop # kafka server uri kafka.server.uri=tcp://localhost:9092 # hdfs location of input directory input=/tmp/kafka/data # hdfs location of output directory output=/tmp/kafka/output # limit the number of events to be fetched; # value -1 means no limitation kafka.request.limit=-1 # kafka parameters client.buffer.size=1048576 client.so.timeout=60000 Any ideas where can I have the problem?