Hi,

I have just compiled kafka from https://github.com/kafka-dev/kafka and
executed the DataGenerator:

./run-class.sh kafka.etl.impl.DataGenerator test/test.properties

After that I have executed the hadoop consumer:

./run-class.sh kafka.etl.impl.SimpleKafkaETLJob test/test.properties


The hadoop-consumer is generating a file on the specified output but it is
never finishing, even if I try to generate only 1 event
at test/test.properties. So this file is growing and growing, my guessing
is that maybe it is reading always the offset 0?

That is my test.properties:

# name of test topic
kafka.etl.topic=SimpleTestEvent5

# hdfs location of jars
hdfs.default.classpath.dir=/tmp/kafka/lib

# number of test events to be generated
event.count=1

# hadoop id and group
hadoop.job.ugi=kafka,hadoop

# kafka server uri
kafka.server.uri=tcp://localhost:9092

# hdfs location of input directory
input=/tmp/kafka/data

# hdfs location of output directory
output=/tmp/kafka/output

# limit the number of events to be fetched;
# value -1 means no limitation
kafka.request.limit=-1

# kafka parameters
client.buffer.size=1048576
client.so.timeout=60000


Any ideas where can I have the problem?

Reply via email to