[
https://issues.apache.org/jira/browse/FLINK-36817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17902975#comment-17902975
]
david radley commented on FLINK-36817:
--------------------------------------
[~lkokhreidze] I had a search in the src tree for _new KafkaConsumer_ . I see
other hits in the code where we create Kafka consumers including
_KafkaPartitionDiscoverer_ and {_}KafkaConsumerThread{_}. I am not sure what
your custom consumer is doing, but I suspect you will want it used everywhere
in place of the default Kafka consumer. WDYT?
> Give users ability to provide their own KafkaConsumer when using
> flink-connector-kafka
> --------------------------------------------------------------------------------------
>
> Key: FLINK-36817
> URL: https://issues.apache.org/jira/browse/FLINK-36817
> Project: Flink
> Issue Type: New Feature
> Components: Connectors / Kafka
> Reporter: Levani Kokhreidze
> Priority: Major
> Labels: pull-request-available
>
> In certain scenarios users of the `KafkaSource` in the flink-connector-kafka
> might want to provide their own KafkaConsumer. Right now this is not possible
> as consumer is created in the
> [KafkaPartitionSplitReader|https://github.com/apache/flink-connector-kafka/blob/main/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/source/reader/KafkaPartitionSplitReader.java#L97]
> which makes customisation impossible.
> Proposal is to let users pass `KafkaConsumerFactory` when building the
> KafkaSource.
> {code:java}
> public interface KafkaConsumerFactory {
> KafkaConsumer<byte[], byte[]> get(Properties properties);
> }{code}
> Builder will have a default implementation which creates the KafkaConsumer
> the same as it is done now.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)