[
https://issues.apache.org/jira/browse/KAFKA-4762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15864954#comment-15864954
]
huxi commented on KAFKA-4762:
-----------------------------
Logs show that you are using 0.10.x where max.partition.fetch.bytes is a hard
limit even when you enable the compression. In your case, seems that you have
enabled the compression on the producer side. `max.partition.fetch.bytes` also
applies to the whole compressed message which is often much larger than a
single one. That's why you run into RecordTooLargeException.
0.10.1 which completes
[KIP-74|https://cwiki.apache.org/confluence/display/KAFKA/KIP-74:+Add+Fetch+Response+Size+Limit+in+Bytes]
already 'fixes' your problem by making `max.partition.fetch.bytes` field in
the fetch request much less useful, so you can try with an 0.10.1 build.
> Consumer throwing RecordTooLargeException even when messages are not that
> large
> -------------------------------------------------------------------------------
>
> Key: KAFKA-4762
> URL: https://issues.apache.org/jira/browse/KAFKA-4762
> Project: Kafka
> Issue Type: Bug
> Reporter: Vipul Singh
>
> We were just recently hit by a weird error.
> Before going in any further, explaining of our service setup. we have a
> producer which produces messages not larger than 256 kb of messages( we have
> an explicit check about this on the producer side) and on the client side we
> have a fetch limit of 512kb(max.partition.fetch.bytes is set to 524288 bytes)
> Recently our client started to see this error:
> {quote}
> org.apache.kafka.common.errors.RecordTooLargeException: There are some
> messages at [Partition=Offset]: {topic_name-0=9925056036} whose size is
> larger than the fetch size 524288 and hence cannot be ever returned. Increase
> the fetch size, or decrease the maximum message size the broker will allow.
> {quote}
> We tried consuming messages with another consumer, without any
> max.partition.fetch.bytes limit, and it consumed fine. The messages were
> small, and did not seem to be greater than 256 kb
> We took a log dump, and the log size looked fine.
> {quote}
> mpresscodec: NoCompressionCodec crc: 2473548911 keysize: 8
> offset: 9925056032 position: 191380053 isvalid: true payloadsize: 539 magic:
> 0 compresscodec: NoCompressionCodec crc: 1656420267 keysize: 8
> offset: 9925056033 position: 191380053 isvalid: true payloadsize: 1551 magic:
> 0 compresscodec: NoCompressionCodec crc: 2398479758 keysize: 8
> offset: 9925056034 position: 191380053 isvalid: true payloadsize: 1307 magic:
> 0 compresscodec: NoCompressionCodec crc: 2845554215 keysize: 8
> offset: 9925056035 position: 191380053 isvalid: true payloadsize: 1520 magic:
> 0 compresscodec: NoCompressionCodec crc: 3106984195 keysize: 8
> offset: 9925056036 position: 191713371 isvalid: true payloadsize: 1207 magic:
> 0 compresscodec: NoCompressionCodec crc: 3462154435 keysize: 8
> offset: 9925056037 position: 191713371 isvalid: true payloadsize: 418 magic:
> 0 compresscodec: NoCompressionCodec crc: 1536701802 keysize: 8
> offset: 9925056038 position: 191713371 isvalid: true payloadsize: 299 magic:
> 0 compresscodec: NoCompressionCodec crc: 4112567543 keysize: 8
> offset: 9925056039 position: 191713371 isvalid: true payloadsize: 1571 magic:
> 0 compresscodec: NoCompressionCodec crc: 3696994307 keysize: 8
> {quote}
> Has anyone seen something similar? or any points to troubleshoot this further
> Please Note: To overcome this issue, we deployed a new consumer, without this
> limit of max.partition.fetch.bytes, and it worked fine.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)