Jun, I did a test this morning and got a very interesting result with you command. I started by wipe all the log files and clean up all zookeeper data files.
Once I restarted both server, producer and consumer then execute your command, what I got is a empty log as following: Dumping /Users/scott/Temp/kafka/test-topic-0/00000000000000000000.log Starting offset: 0 One observation, the 00000000000000000000.index file was getting huge but there was nothing in 00000000000000000000.log file. Thanks, Scott On Tue, Jul 9, 2013 at 8:40 PM, Jun Rao <jun...@gmail.com> wrote: > Could you run the following command on one of the log files of your topic > and attach the output? > > bin/kafka-run-class.sh kafka.tools.DumpLogSegments --files > /tmp/kafka-logs/testtopic-0/00000000000000000000.log > > Thanks, > > Jun > > > On Tue, Jul 9, 2013 at 3:23 PM, Scott Wang < > scott.w...@rumbleentertainment.com> wrote: > > > Another piece of information, the snappy compression also does not work. > > > > Thanks, > > Scott > > > > > > On Tue, Jul 9, 2013 at 11:07 AM, Scott Wang < > > scott.w...@rumbleentertainment.com> wrote: > > > > > I just try it and it still not showing up, thanks for looking into > this. > > > > > > Thanks, > > > Scott > > > > > > > > > On Tue, Jul 9, 2013 at 8:06 AM, Jun Rao <jun...@gmail.com> wrote: > > > > > >> Could you try starting the consumer first (and enable gzip in the > > >> producer)? > > >> > > >> Thanks, > > >> > > >> Jun > > >> > > >> > > >> On Mon, Jul 8, 2013 at 9:37 PM, Scott Wang < > > >> scott.w...@rumbleentertainment.com> wrote: > > >> > > >> > No, I did not start the consumer before the producer. I actually > > >> started > > >> > the producer first and nothing showed up in the consumer unless I > > >> commented > > >> > out this line -- props.put("compression.codec", "gzip"). If I > > >> commented > > >> > out the compression codec, everything just works. > > >> > > > >> > > > >> > On Mon, Jul 8, 2013 at 9:07 PM, Jun Rao <jun...@gmail.com> wrote: > > >> > > > >> > > Did you start the consumer before the producer? Be default, the > > >> consumer > > >> > > gets only the new data? > > >> > > > > >> > > Thanks, > > >> > > > > >> > > Jun > > >> > > > > >> > > > > >> > > On Mon, Jul 8, 2013 at 2:53 PM, Scott Wang < > > >> > > scott.w...@rumbleentertainment.com> wrote: > > >> > > > > >> > > > I am testing with Kafka 0.8 beta and having problem of receiving > > >> > message > > >> > > in > > >> > > > consumer. There is no error so does anyone have any insights. > > >> When I > > >> > > > commented out the "compression.code" everything works fine. > > >> > > > > > >> > > > My producer: > > >> > > > public class TestKafka08Prod { > > >> > > > > > >> > > > public static void main(String [] args) { > > >> > > > > > >> > > > Producer<Integer, String> producer = null; > > >> > > > try { > > >> > > > Properties props = new Properties(); > > >> > > > props.put("metadata.broker.list", "localhost:9092"); > > >> > > > props.put("serializer.class", > > >> > > > "kafka.serializer.StringEncoder"); > > >> > > > props.put("producer.type", "sync"); > > >> > > > props.put("request.required.acks","1"); > > >> > > > props.put("compression.codec", "gzip"); > > >> > > > ProducerConfig config = new ProducerConfig(props); > > >> > > > producer = new Producer<Integer, String>(config); > > >> > > > int j=0; > > >> > > > for(int i=0; i<10; i++) { > > >> > > > KeyedMessage<Integer, String> data = new > > >> > > > KeyedMessage<Integer, String>("test-topic", "test-message: "+i+" > > >> > > > "+System.currentTimeMillis()); > > >> > > > producer.send(data); > > >> > > > > > >> > > > } > > >> > > > > > >> > > > } catch (Exception e) { > > >> > > > System.out.println("Error happened: "); > > >> > > > e.printStackTrace(); > > >> > > > } finally { > > >> > > > if(null != null) { > > >> > > > producer.close(); > > >> > > > } > > >> > > > > > >> > > > System.out.println("Ened of Sending"); > > >> > > > } > > >> > > > > > >> > > > System.exit(0); > > >> > > > } > > >> > > > } > > >> > > > > > >> > > > > > >> > > > My consumer: > > >> > > > > > >> > > > public class TestKafka08Consumer { > > >> > > > public static void main(String [] args) throws > > >> > UnknownHostException, > > >> > > > SocketException { > > >> > > > > > >> > > > Properties props = new Properties(); > > >> > > > props.put("zookeeper.connect", > > "localhost:2181/kafka_0_8"); > > >> > > > props.put("group.id", "test08ConsumerId"); > > >> > > > props.put("zk.sessiontimeout.ms", "4000"); > > >> > > > props.put("zk.synctime.ms", "2000"); > > >> > > > props.put("autocommit.interval.ms", "1000"); > > >> > > > > > >> > > > ConsumerConfig consumerConfig = new > ConsumerConfig(props); > > >> > > > > > >> > > > ConsumerConnector consumerConnector = > > >> > > > > > kafka.consumer.Consumer.createJavaConsumerConnector(consumerConfig); > > >> > > > > > >> > > > String topic = "test-topic"; > > >> > > > Map<String, Integer> topicCountMap = new HashMap<String, > > >> > > > Integer>(); > > >> > > > topicCountMap.put(topic, new Integer(1)); > > >> > > > Map<String, List<KafkaStream<byte[], byte[]>>> > > consumerMap = > > >> > > > consumerConnector.createMessageStreams(topicCountMap); > > >> > > > KafkaStream<byte[], byte[]> stream = > > >> > > > consumerMap.get(topic).get(0); > > >> > > > > > >> > > > ConsumerIterator<byte[], byte[]> it = stream.iterator(); > > >> > > > > > >> > > > int counter=0; > > >> > > > while(it.hasNext()) { > > >> > > > try { > > >> > > > String fromPlatform = new > > >> String(it.next().message()); > > >> > > > System.out.println("The messages: > "+fromPlatform); > > >> > > > } catch(Exception e) { > > >> > > > e.printStackTrace(); > > >> > > > } > > >> > > > } > > >> > > > System.out.println("SystemOut"); > > >> > > > } > > >> > > > } > > >> > > > > > >> > > > > > >> > > > Thanks > > >> > > > > > >> > > > > >> > > > >> > > > > > > > > >