Brian Byrne created KAFKA-8904:
----------------------------------
Summary: Reduce metadata lookups when producting to a large number
of topics
Key: KAFKA-8904
URL: https://issues.apache.org/jira/browse/KAFKA-8904
Project: Kafka
Issue Type: Improvement
Components: controller, producer
Reporter: Brian Byrne
Per [~lbradstreet]:
"The problem was that the producer starts with no knowledge of topic metadata.
So they start the producer up, and then they start sending messages to any of
the thousands of topics that exist. Each time a message is sent to a new topic,
it'll trigger a metadata request if the producer doesn't know about it. These
metadata requests are done in serial such that if you send 2000 messages to
2000 topics, it will trigger 2000 new metadata requests.
Each successive metadata request will include every topic seen so far, so the
first metadata request will include 1 topic, the second will include 2 topics,
etc.
An additional problem is that this can take a while, and metadata expiry (for
metadata that has not been recently used) is hard coded to 5 mins, so if this
the initial fetches take long enough you can end up evicting the metadata
before you send another message to a topic.
So the approaches above are:
1. We can linger for a bit before making a metadata request, allow more sends
to go through, and then batch the metadata request for topics we we need in a
single metadata request.
2. We can allow pre-seeding the producer with metadata for a list of topics you
care about.
I prefer 1 if we can make it work."
--
This message was sent by Atlassian Jira
(v8.3.2#803003)