Jun,

I did a test this morning and got a very interesting result with you
command.  I started by wipe all the log files and clean up all zookeeper
data files.

Once I restarted both server, producer and consumer then execute your
command, what I got is a empty log as following:

Dumping /Users/scott/Temp/kafka/test-topic-0/00000000000000000000.log
Starting offset: 0

One observation, the 00000000000000000000.index file was getting huge but
there was nothing in 00000000000000000000.log file.

Thanks,
Scott




On Tue, Jul 9, 2013 at 8:40 PM, Jun Rao <jun...@gmail.com> wrote:

> Could you run the following command on one of the log files of your topic
> and attach the output?
>
> bin/kafka-run-class.sh kafka.tools.DumpLogSegments --files
> /tmp/kafka-logs/testtopic-0/00000000000000000000.log
>
> Thanks,
>
> Jun
>
>
> On Tue, Jul 9, 2013 at 3:23 PM, Scott Wang <
> scott.w...@rumbleentertainment.com> wrote:
>
> > Another piece of information, the snappy compression also does not work.
> >
> > Thanks,
> > Scott
> >
> >
> > On Tue, Jul 9, 2013 at 11:07 AM, Scott Wang <
> > scott.w...@rumbleentertainment.com> wrote:
> >
> > > I just try it and it still not showing up, thanks for looking into
> this.
> > >
> > > Thanks,
> > > Scott
> > >
> > >
> > > On Tue, Jul 9, 2013 at 8:06 AM, Jun Rao <jun...@gmail.com> wrote:
> > >
> > >> Could you try starting the consumer first (and enable gzip in the
> > >> producer)?
> > >>
> > >> Thanks,
> > >>
> > >> Jun
> > >>
> > >>
> > >> On Mon, Jul 8, 2013 at 9:37 PM, Scott Wang <
> > >> scott.w...@rumbleentertainment.com> wrote:
> > >>
> > >> > No, I did not start the consumer before the producer.  I actually
> > >> started
> > >> > the producer first and nothing showed up in the consumer unless I
> > >> commented
> > >> > out this line -- props.put("compression.codec", "gzip").    If I
> > >> commented
> > >> > out the compression codec, everything just works.
> > >> >
> > >> >
> > >> > On Mon, Jul 8, 2013 at 9:07 PM, Jun Rao <jun...@gmail.com> wrote:
> > >> >
> > >> > > Did you start the consumer before the producer? Be default, the
> > >> consumer
> > >> > > gets only the new data?
> > >> > >
> > >> > > Thanks,
> > >> > >
> > >> > > Jun
> > >> > >
> > >> > >
> > >> > > On Mon, Jul 8, 2013 at 2:53 PM, Scott Wang <
> > >> > > scott.w...@rumbleentertainment.com> wrote:
> > >> > >
> > >> > > > I am testing with Kafka 0.8 beta and having problem of receiving
> > >> > message
> > >> > > in
> > >> > > > consumer.  There is no error so does anyone have any insights.
> > >>  When I
> > >> > > > commented out the "compression.code" everything works fine.
> > >> > > >
> > >> > > > My producer:
> > >> > > > public class TestKafka08Prod {
> > >> > > >
> > >> > > >     public static void main(String [] args) {
> > >> > > >
> > >> > > >         Producer<Integer, String> producer = null;
> > >> > > >         try {
> > >> > > >             Properties props = new Properties();
> > >> > > >             props.put("metadata.broker.list", "localhost:9092");
> > >> > > >             props.put("serializer.class",
> > >> > > > "kafka.serializer.StringEncoder");
> > >> > > >             props.put("producer.type", "sync");
> > >> > > >             props.put("request.required.acks","1");
> > >> > > >             props.put("compression.codec", "gzip");
> > >> > > >             ProducerConfig config = new ProducerConfig(props);
> > >> > > >             producer = new Producer<Integer, String>(config);
> > >> > > >             int j=0;
> > >> > > >             for(int i=0; i<10; i++) {
> > >> > > >                 KeyedMessage<Integer, String> data = new
> > >> > > > KeyedMessage<Integer, String>("test-topic", "test-message: "+i+"
> > >> > > > "+System.currentTimeMillis());
> > >> > > >                 producer.send(data);
> > >> > > >
> > >> > > >             }
> > >> > > >
> > >> > > >         } catch (Exception e) {
> > >> > > >             System.out.println("Error happened: ");
> > >> > > >             e.printStackTrace();
> > >> > > >         } finally {
> > >> > > >             if(null != null) {
> > >> > > >                 producer.close();
> > >> > > >             }
> > >> > > >
> > >> > > >             System.out.println("Ened of Sending");
> > >> > > >         }
> > >> > > >
> > >> > > >         System.exit(0);
> > >> > > >     }
> > >> > > > }
> > >> > > >
> > >> > > >
> > >> > > > My consumer:
> > >> > > >
> > >> > > > public class TestKafka08Consumer {
> > >> > > >     public static void main(String [] args) throws
> > >> > UnknownHostException,
> > >> > > > SocketException {
> > >> > > >
> > >> > > >         Properties props = new Properties();
> > >> > > >         props.put("zookeeper.connect",
> > "localhost:2181/kafka_0_8");
> > >> > > >         props.put("group.id", "test08ConsumerId");
> > >> > > >         props.put("zk.sessiontimeout.ms", "4000");
> > >> > > >         props.put("zk.synctime.ms", "2000");
> > >> > > >         props.put("autocommit.interval.ms", "1000");
> > >> > > >
> > >> > > >         ConsumerConfig consumerConfig = new
> ConsumerConfig(props);
> > >> > > >
> > >> > > >         ConsumerConnector consumerConnector =
> > >> > > >
> > kafka.consumer.Consumer.createJavaConsumerConnector(consumerConfig);
> > >> > > >
> > >> > > >         String topic = "test-topic";
> > >> > > >         Map<String, Integer> topicCountMap = new HashMap<String,
> > >> > > > Integer>();
> > >> > > >         topicCountMap.put(topic, new Integer(1));
> > >> > > >         Map<String, List<KafkaStream<byte[], byte[]>>>
> > consumerMap =
> > >> > > > consumerConnector.createMessageStreams(topicCountMap);
> > >> > > >         KafkaStream<byte[], byte[]> stream =
> > >> > > >  consumerMap.get(topic).get(0);
> > >> > > >
> > >> > > >         ConsumerIterator<byte[], byte[]> it = stream.iterator();
> > >> > > >
> > >> > > >         int counter=0;
> > >> > > >         while(it.hasNext()) {
> > >> > > >             try {
> > >> > > >                 String fromPlatform = new
> > >> String(it.next().message());
> > >> > > >                 System.out.println("The messages:
> "+fromPlatform);
> > >> > > >             } catch(Exception e) {
> > >> > > >                 e.printStackTrace();
> > >> > > >             }
> > >> > > >         }
> > >> > > >         System.out.println("SystemOut");
> > >> > > >     }
> > >> > > > }
> > >> > > >
> > >> > > >
> > >> > > > Thanks
> > >> > > >
> > >> > >
> > >> >
> > >>
> > >
> > >
> >
>

Reply via email to