Ruben Ramalho created SPARK-9476:
------------------------------------

             Summary: Kafka stream loses leader after 2h of operation 
                 Key: SPARK-9476
                 URL: https://issues.apache.org/jira/browse/SPARK-9476
             Project: Spark
          Issue Type: Bug
          Components: Streaming
    Affects Versions: 1.4.1
         Environment: Docker, Centos, Spark standalone, core i7, 8Gb
            Reporter: Ruben Ramalho


This seems to happen every 2h, it happens both with the direct stream and 
regular stream, I'm doing window operations over a 1h period (if that can help).

Here's part of the error message:

2015-07-30 13:27:23 WARN  ClientUtils$:89 - Fetching topic metadata with 
correlation id 10 for topics [Set(updates)] from broker 
[id:0,host:192.168.3.23,port:3000] failed
java.nio.channels.ClosedChannelException
        at kafka.network.BlockingChannel.send(BlockingChannel.scala:100)
        at kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:73)
        at 
kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:72)
        at kafka.producer.SyncProducer.send(SyncProducer.scala:113)
        at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:58)
        at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:93)
        at 
kafka.consumer.ConsumerFetcherManager$LeaderFinderThread.doWork(ConsumerFetcherManager.scala:66)
        at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60)
2015-07-30 13:27:23 INFO  SyncProducer:68 - Disconnecting from 192.168.3.23:3000
2015-07-30 13:27:23 WARN  ConsumerFetcherManager$LeaderFinderThread:89 - 
[spark-group_81563e123e9f-1438259236988-fc3d82bf-leader-finder-thread], Failed 
to find leader for Set([updates,0])
kafka.common.KafkaException: fetching topic metadata for topics 
[Set(oversight-updates)] from broker 
[ArrayBuffer(id:0,host:192.168.3.23,port:3000)] failed
        at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:72)
        at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:93)
        at 
kafka.consumer.ConsumerFetcherManager$LeaderFinderThread.doWork(ConsumerFetcherManager.scala:66)
        at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60)
Caused by: java.nio.channels.ClosedChannelException
        at kafka.network.BlockingChannel.send(BlockingChannel.scala:100)
        at kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:73)
        at 
kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:72)
        at kafka.producer.SyncProducer.send(SyncProducer.scala:113)
        at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:58)

After the crash I tried to communicate with kafka with a simple scala consumer 
and producer and have no problem at all. Spark tough needs a kafka container 
restart to start normal operaiton. There are no errors on the kafka log, apart 
from an improper closed connection.

I have been trying to solve this problem for days, I suspect this has something 
to do with spark.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to