karthickthavasiraj09 created KAFKA-19238:
--------------------------------------------
Summary: We're facing issue in Kafka while reading data from Azure
event hubs through Azure Databricks
Key: KAFKA-19238
URL: https://issues.apache.org/jira/browse/KAFKA-19238
Project: Kafka
Issue Type: Test
Components: connect, consumer, network
Affects Versions: 3.3.1
Environment: Production
Reporter: karthickthavasiraj09
Attachments: Re_ Job taking much longer time to extract data... -
TrackingID#2504010030002895.msg
We had an issue while reading data from the Azure Event hubs through Azure
Databricks. After working with Microsoft team they confirmed that there's an
issue from Kafka side. I'm sharing the debug logs shared by the Microsoft team
below,
The good job shared on March 20th, so we would not be able to download the
backend logs _(as it's > 20 days)_
But for the bad job:
[https://adb-2632737963103362.2.azuredatabricks.net/jobs/911028616577296/runs/939144212532710?o=2632737963103362]
that took 49m, we see that task 143 takes 46 mins _(out of the job duration of_
_49m 30s)_
_25/04/15 14:21:44 INFO KafkaBatchReaderFactoryWithRowBytesAccumulator:
Creating Kafka reader
topicPartition=voyager-prod-managedsql-cus.order.orders.orderitem-0
fromOffset=16511904 untilOffset=16658164, for query
queryId=dd660d4d-05cc-4a8e-8f93-d202ec78fec3
runId=af7eb711-7310-4788-85b7-0977fc0756b7 batchId=73 taskId=143 partitionId=0_
_._
_25/04/15 15:07:21 INFO KafkaDataConsumer: From Kafka
topicPartition=voyager-prod-managedsql-cus.order.orders.orderitem-0
groupId=spark-kafka-source-da79e0fc-8ee5-40f5-a127-7b31766b3022--1737876659-executor
read 146260 records through 4314 polls (polled out 146265 records), taking
2526471821132 nanos, during time span of 2736294068630 nanos._
And this task is waiting for Kafka to respond for most of the time as we can
see from the threads:
_Executor task launch worker for task 0.0 in stage 147.0 (TID 143)_
_sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)_
_sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)_
_sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93)_
_sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) - locked
sun.nio.ch.EPollSelectorImpl@54f8f9b6_
_sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)_
_kafkashaded.org.apache.kafka.common.network.Selector.select(Selector.java:874)_
_kafkashaded.org.apache.kafka.common.network.Selector.poll(Selector.java:465)_
_kafkashaded.org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:560)_
_kafkashaded.org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:280)_
_kafkashaded.org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:251)_
_kafkashaded.org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:242)_
_kafkashaded.org.apache.kafka.clients.consumer.KafkaConsumer.position(KafkaConsumer.java:1759)_
_kafkashaded.org.apache.kafka.clients.consumer.KafkaConsumer.position(KafkaConsumer.java:1717)_
_org.apache.spark.sql.kafka010.consumer.InternalKafkaConsumer.getAvailableOffsetRange(KafkaDataConsumer.scala:110)_
_org.apache.spark.sql.kafka010.consumer.InternalKafkaConsumer.fetch(KafkaDataConsumer.scala:84)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.$anonfun$fetchData$1(KafkaDataConsumer.scala:593)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer$$Lambda$4556/228899458.apply(Unknown
Source)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.timeNanos(KafkaDataConsumer.scala:696)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.fetchData(KafkaDataConsumer.scala:593)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.fetchRecord(KafkaDataConsumer.scala:517)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.$anonfun$get$1(KafkaDataConsumer.scala:325)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer$$Lambda$4491/342980175.apply(Unknown
Source)_
_org.apache.spark.util.UninterruptibleThread.runUninterruptibly(UninterruptibleThread.scala:77)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.runUninterruptiblyIfPossible(KafkaDataConsumer.scala:686)_
_org.apache.spark.sql.kafka010.consumer.KafkaDataConsumer.get(KafkaDataConsumer.scala:301)_
_org.apache.spark.sql.kafka010.KafkaBatchPartitionReader.next(KafkaBatchPartitionReader.scala:106)_
_._
--
This message was sent by Atlassian Jira
(v8.20.10#820010)