[
https://issues.apache.org/jira/browse/KAFKA-3995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15895443#comment-15895443
]
ASF GitHub Bot commented on KAFKA-3995:
---------------------------------------
GitHub user becketqin opened a pull request:
https://github.com/apache/kafka/pull/2638
KAFKA-3995: fix compression ratio estimation.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/becketqin/kafka KAFKA-3995
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/kafka/pull/2638.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #2638
----
commit cee4d58cb0afaa526bddd372246894bab82c82c2
Author: Jiangjie Qin <[email protected]>
Date: 2017-03-04T01:40:12Z
KAFKA-3995: fix compression ratio estimation.
----
> Add a new configuration "enable.compression.ratio.estimation" to the producer
> config
> ------------------------------------------------------------------------------------
>
> Key: KAFKA-3995
> URL: https://issues.apache.org/jira/browse/KAFKA-3995
> Project: Kafka
> Issue Type: Improvement
> Components: clients
> Affects Versions: 0.10.0.0
> Reporter: Jiangjie Qin
> Assignee: Mayuresh Gharat
>
> We recently see a few cases where RecordTooLargeException is thrown because
> the compressed message sent by KafkaProducer exceeded the max message size.
> The root cause of this issue is because the compressor is estimating the
> batch size using an estimated compression ratio based on heuristic
> compression ratio statistics. This does not quite work for the traffic with
> highly variable compression ratios.
> For example, if the batch size is set to 1MB and the max message size is 1MB.
> Initially a the producer is sending messages (each message is 1MB) to topic_1
> whose data can be compressed to 1/10 of the original size. After a while the
> estimated compression ratio in the compressor will be trained to 1/10 and the
> producer would put 10 messages into one batch. Now the producer starts to
> send messages (each message is also 1MB) to topic_2 whose message can only be
> compress to 1/5 of the original size. The producer would still use 1/10 as
> the estimated compression ratio and put 10 messages into a batch. That batch
> would be 2 MB after compression which exceeds the maximum message size. In
> this case the user do not have many options other than resend everything or
> close the producer if they care about ordering.
> This is especially an issue for services like MirrorMaker whose producer is
> shared by many different topics.
> To solve this issue, we can probably add a configuration
> "enable.compression.ratio.estimation" to the producer. So when this
> configuration is set to false, we stop estimating the compressed size but
> will close the batch once the uncompressed bytes in the batch reaches the
> batch size.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)