[ 
https://issues.apache.org/jira/browse/KAFKA-19403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17966775#comment-17966775
 ] 

Matthias J. Sax commented on KAFKA-19403:
-----------------------------------------

The provide log snippets do not appear to be Kafka related. There is no classes 
`KafkaBatchReaderFactoryWithRowBytesAccumulator` or `KafkaDataConsumer` in the 
Kafka code. As a matter of fact, it seems `KafkaDataConsumer` is from Spark (-> 
`org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer`)

At this point, it's totally unclear if there is actually a bug or if this is a 
configuration issue, or Event Hubs, or Databricks issue. – Last but not least, 
3.3.1 is an older Kafka versions, so the main question would be, if you can 
reproduce the issue in Kafka 4.0.0?
{quote}After conducting initial analysis with the Microsoft support team, they 
confirmed that the root cause appears to be related to Kafka
{quote}
This is very vague, and we won't be able to help with a more concrete problem 
statement, or other information. How can this issue be reproduced? Is there any 
broker / client side ERROR logs or similar that indicate a bug?

If you suspect a issue in client-broker communication, the first step would be 
to see if the client did send fetch request to the broker, and if the broker 
did receive and process these request. Client/broker side debug (or trace) 
level logs should contain this information.

> We're facing issue in Kafka while reading data from Azure event hubs through 
> Azure Databricks
> ---------------------------------------------------------------------------------------------
>
>                 Key: KAFKA-19403
>                 URL: https://issues.apache.org/jira/browse/KAFKA-19403
>             Project: Kafka
>          Issue Type: Bug
>          Components: consumer, network
>    Affects Versions: 3.3.1
>         Environment: Production
>            Reporter: karthickthavasiraj09
>            Priority: Major
>
> * We are experiencing significant slowness while reading data from Azure 
> Event Hubs using Azure Databricks. After conducting initial analysis with the 
> Microsoft support team, they confirmed that the root cause appears to be 
> related to Kafka. We are reaching out for your assistance in investigating 
> and resolving this issue.
> Below are the key findings and debug logs provided by the Microsoft team:
>  ** The data read operation took *49 minutes* in total.
>  ** Out of this, only one task(spark task) {*}Task 143 alone took 46 
> minutes{*}, indicating a bottleneck in this specific task.
>  ** The job duration was {*}49 minutes and 30 seconds{*}.
> Relevant Log Snippets:
> 25/04/15 14:21:44 INFO KafkaBatchReaderFactoryWithRowBytesAccumulator:
> Creating Kafka reader topicPartition=<topic-name>-0 fromOffset=16511904 
> untilOffset=16658164,
> for queryId=dd660d4d-05cc-4a8e-8f93-d202ec78fec3 
> runId=af7eb711-7310-4788-85b7-0977fc0756b7 batchId=73 taskId=143 partitionId=0
> 25/04/15 15:07:21 INFO KafkaDataConsumer:
> From Kafka topicPartition=<topic-name>-0 
> groupId=spark-kafka-source-da79e0fc-8ee5-40f5-a127-7b31766b3022--1737876659-executor
> read 146260 records through 4314 polls (polled out 146265 records), taking 
> 2526471821132 ns,
> over a timespan of 2736294068630 ns.
>  
> Additionally, the thread stack trace indicates that the task was mostly 
> waiting on Kafka to respond. See the following thread details captured during 
> the slowness:
> Executor task launch worker for task 0.0 in stage 147.0 (TID 143)
> sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
> ...
> kafkashaded.org.apache.kafka.clients.consumer.KafkaConsumer.position(KafkaConsumer.java:1759)
> ...
> org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.fetchRecord(KafkaDataConsumer.scala:517)
> *We kindly request the Kafka team to look into this issue. The task appears 
> to be blocked for an extended period during Kafka polling. Any insights into 
> why Kafka is not responding promptly or recommendations for configuration 
> changes or optimizations would be greatly appreciated.*
> *Please let us know if any additional information or diagnostics are required 
> from our end.*
> *Thank you for your support.*



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to