If you increase the size of the messages for producing then you **MUST** also
change *replica.fetch.max.bytes i*n the broker* server.properties *otherwise
none of your replicas will be able to fetch from the leader and they will
all fall out of the ISR. You also then need to change your consumers
*fetch.message.max.bytes* in your consumers properties (whoever that might
be configured for your specific consumer being used) so that they can read
that data otherwise you won't see messages downstream.

/*******************************************
 Joe Stein
 Founder, Principal Consultant
 Big Data Open Source Security LLC
 http://www.stealth.ly
 Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
********************************************/

On Mon, Jan 19, 2015 at 1:03 PM, Magnus Edenhill <mag...@edenhill.se> wrote:

> (duplicating the github answer for reference)
>
> Hi Eduardo,
>
> the default maximum fetch size is 1 Meg which means your 2 Meg messages
> will not fit the fetch request.
> Try increasing it by appending -X fetch.message.max.bytes=4000000 to your
> command line.
>
> Regards,
> Magnus
>
>
> 2015-01-19 17:52 GMT+01:00 Eduardo Costa Alfaia <e.costaalf...@unibs.it>:
>
> > Hi All,
> > I am having an issue when using kafka with librdkafka. I've changed the
> > message.max.bytes to 2MB in my server.properties config file, that is the
> > size of my message, when I run the command line ./rdkafka_performance -C
> -t
> > test -p 0 -b computer49:9092, after consume some messages the consumer
> > remain waiting something that don't arrive. My producer continues sending
> > messages. Some idea?
> >
> > % Using random seed 1421685059, verbosity level 1
> > % 214 messages and 1042835 bytes consumed in 20ms: 10518 msgs/s and 51.26
> > Mb/s, no compression
> > % 21788 messages and 106128192 bytes consumed in 1029ms: 21154 msgs/s and
> > 103.04 Mb/s, no compression
> > % 43151 messages and 210185259 bytes consumed in 2030ms: 21252 msgs/s and
> > 103.52 Mb/s, no compression
> > % 64512 messages and 314233575 bytes consumed in 3031ms: 21280 msgs/s and
> > 103.66 Mb/s, no compression
> > % 86088 messages and 419328692 bytes consumed in 4039ms: 21313 msgs/s and
> > 103.82 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 5719ms: 17571 msgs/s
> and
> > 85.67 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 6720ms: 14955 msgs/s
> and
> > 72.92 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 7720ms: 13018 msgs/s
> and
> > 63.47 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 8720ms: 11524 msgs/s
> and
> > 56.19 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 9720ms: 10339 msgs/s
> and
> > 50.41 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 10721ms: 9374 msgs/s
> and
> > 45.71 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 11721ms: 8574 msgs/s
> and
> > 41.81 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 12721ms: 7900 msgs/s
> and
> > 38.52 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 13721ms: 7324 msgs/s
> and
> > 35.71 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 14721ms: 6826 msgs/s
> and
> > 33.29 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 15722ms: 6392 msgs/s
> and
> > 31.17 Mb/s, no compression
> > % 100504 messages and 490022646 bytes consumed in 16722ms: 6010 msgs/s
> and
> > 29.30 Mb/s, no
> > ........
> >
> >
> > The software when consume all offset send me the message:
> >
> > % Consumer reached end of unibs.nec [0] message queue at offset 229790
> > RD_KAFKA_RESP_ERR__PARTITION_EOF: [-191]
> >
> > However changed de message.max.bytes to 2MB I don’t receive the code from
> > Kafka.
> >
> > Anyone has some idea?
> >
> > Thanks guys.
> > --
> > Informativa sulla Privacy: http://www.unibs.it/node/8155
> >
>

Reply via email to