Could you run the following command on one of the log files of your topic
and attach the output?

bin/kafka-run-class.sh kafka.tools.DumpLogSegments --files
/tmp/kafka-logs/testtopic-0/00000000000000000000.log

Thanks,

Jun


On Tue, Jul 9, 2013 at 3:23 PM, Scott Wang <
scott.w...@rumbleentertainment.com> wrote:

> Another piece of information, the snappy compression also does not work.
>
> Thanks,
> Scott
>
>
> On Tue, Jul 9, 2013 at 11:07 AM, Scott Wang <
> scott.w...@rumbleentertainment.com> wrote:
>
> > I just try it and it still not showing up, thanks for looking into this.
> >
> > Thanks,
> > Scott
> >
> >
> > On Tue, Jul 9, 2013 at 8:06 AM, Jun Rao <jun...@gmail.com> wrote:
> >
> >> Could you try starting the consumer first (and enable gzip in the
> >> producer)?
> >>
> >> Thanks,
> >>
> >> Jun
> >>
> >>
> >> On Mon, Jul 8, 2013 at 9:37 PM, Scott Wang <
> >> scott.w...@rumbleentertainment.com> wrote:
> >>
> >> > No, I did not start the consumer before the producer.  I actually
> >> started
> >> > the producer first and nothing showed up in the consumer unless I
> >> commented
> >> > out this line -- props.put("compression.codec", "gzip").    If I
> >> commented
> >> > out the compression codec, everything just works.
> >> >
> >> >
> >> > On Mon, Jul 8, 2013 at 9:07 PM, Jun Rao <jun...@gmail.com> wrote:
> >> >
> >> > > Did you start the consumer before the producer? Be default, the
> >> consumer
> >> > > gets only the new data?
> >> > >
> >> > > Thanks,
> >> > >
> >> > > Jun
> >> > >
> >> > >
> >> > > On Mon, Jul 8, 2013 at 2:53 PM, Scott Wang <
> >> > > scott.w...@rumbleentertainment.com> wrote:
> >> > >
> >> > > > I am testing with Kafka 0.8 beta and having problem of receiving
> >> > message
> >> > > in
> >> > > > consumer.  There is no error so does anyone have any insights.
> >>  When I
> >> > > > commented out the "compression.code" everything works fine.
> >> > > >
> >> > > > My producer:
> >> > > > public class TestKafka08Prod {
> >> > > >
> >> > > >     public static void main(String [] args) {
> >> > > >
> >> > > >         Producer<Integer, String> producer = null;
> >> > > >         try {
> >> > > >             Properties props = new Properties();
> >> > > >             props.put("metadata.broker.list", "localhost:9092");
> >> > > >             props.put("serializer.class",
> >> > > > "kafka.serializer.StringEncoder");
> >> > > >             props.put("producer.type", "sync");
> >> > > >             props.put("request.required.acks","1");
> >> > > >             props.put("compression.codec", "gzip");
> >> > > >             ProducerConfig config = new ProducerConfig(props);
> >> > > >             producer = new Producer<Integer, String>(config);
> >> > > >             int j=0;
> >> > > >             for(int i=0; i<10; i++) {
> >> > > >                 KeyedMessage<Integer, String> data = new
> >> > > > KeyedMessage<Integer, String>("test-topic", "test-message: "+i+"
> >> > > > "+System.currentTimeMillis());
> >> > > >                 producer.send(data);
> >> > > >
> >> > > >             }
> >> > > >
> >> > > >         } catch (Exception e) {
> >> > > >             System.out.println("Error happened: ");
> >> > > >             e.printStackTrace();
> >> > > >         } finally {
> >> > > >             if(null != null) {
> >> > > >                 producer.close();
> >> > > >             }
> >> > > >
> >> > > >             System.out.println("Ened of Sending");
> >> > > >         }
> >> > > >
> >> > > >         System.exit(0);
> >> > > >     }
> >> > > > }
> >> > > >
> >> > > >
> >> > > > My consumer:
> >> > > >
> >> > > > public class TestKafka08Consumer {
> >> > > >     public static void main(String [] args) throws
> >> > UnknownHostException,
> >> > > > SocketException {
> >> > > >
> >> > > >         Properties props = new Properties();
> >> > > >         props.put("zookeeper.connect",
> "localhost:2181/kafka_0_8");
> >> > > >         props.put("group.id", "test08ConsumerId");
> >> > > >         props.put("zk.sessiontimeout.ms", "4000");
> >> > > >         props.put("zk.synctime.ms", "2000");
> >> > > >         props.put("autocommit.interval.ms", "1000");
> >> > > >
> >> > > >         ConsumerConfig consumerConfig = new ConsumerConfig(props);
> >> > > >
> >> > > >         ConsumerConnector consumerConnector =
> >> > > >
> kafka.consumer.Consumer.createJavaConsumerConnector(consumerConfig);
> >> > > >
> >> > > >         String topic = "test-topic";
> >> > > >         Map<String, Integer> topicCountMap = new HashMap<String,
> >> > > > Integer>();
> >> > > >         topicCountMap.put(topic, new Integer(1));
> >> > > >         Map<String, List<KafkaStream<byte[], byte[]>>>
> consumerMap =
> >> > > > consumerConnector.createMessageStreams(topicCountMap);
> >> > > >         KafkaStream<byte[], byte[]> stream =
> >> > > >  consumerMap.get(topic).get(0);
> >> > > >
> >> > > >         ConsumerIterator<byte[], byte[]> it = stream.iterator();
> >> > > >
> >> > > >         int counter=0;
> >> > > >         while(it.hasNext()) {
> >> > > >             try {
> >> > > >                 String fromPlatform = new
> >> String(it.next().message());
> >> > > >                 System.out.println("The messages: "+fromPlatform);
> >> > > >             } catch(Exception e) {
> >> > > >                 e.printStackTrace();
> >> > > >             }
> >> > > >         }
> >> > > >         System.out.println("SystemOut");
> >> > > >     }
> >> > > > }
> >> > > >
> >> > > >
> >> > > > Thanks
> >> > > >
> >> > >
> >> >
> >>
> >
> >
>

Reply via email to