[
https://issues.apache.org/jira/browse/SPARK-21836?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
yue long updated SPARK-21836:
-----------------------------
Description:
When using the package spark-streaming-kafka-0-8 for accessing kafka in spark
dstream, many user will face the "could not find leader" exception if some of
kafka brokers are down. This will cause the whole streaming fail, like
[SPARK-18983|https://issues.apache.org/jira/browse/SPARK-18983] said. The
failed kafka brokers may also cause other problems when creating Dstream or
creating the batch job.
Even though the down of kafka broker is not the bug of spark streaming, we can
avoid this failure in spark streaming. Especially for the reasons that kakfa
cluster is not always stable in the real production.
Does anyoner think we should add some retry logic when kakfa broker is down? I
have implement this function in spark 1.6.3 and spark 2.1.0, and test them. If
we implement this function, it will reduce the failure number of
kafka-streaming which may help streaming users.
was:When using the package spark-streaming-kafka-0-8 for accessing kafka in
spark dstream, many user will face the "could not find leader" exception if
some of kafka brokers are down. This will cause the whole streaming fail. Like
[SPARK-18983|https://issues.apache.org/jira/browse/SPARK-18983] said. The
failed kafka brokers may also cause other problems like creating Dstream or
creating
> [STREAMING] Retry when kafka broker is down in kafka-streaming-0-8
> ------------------------------------------------------------------
>
> Key: SPARK-21836
> URL: https://issues.apache.org/jira/browse/SPARK-21836
> Project: Spark
> Issue Type: Improvement
> Components: DStreams
> Affects Versions: 1.6.3, 2.1.0
> Reporter: yue long
>
> When using the package spark-streaming-kafka-0-8 for accessing kafka in spark
> dstream, many user will face the "could not find leader" exception if some of
> kafka brokers are down. This will cause the whole streaming fail, like
> [SPARK-18983|https://issues.apache.org/jira/browse/SPARK-18983] said. The
> failed kafka brokers may also cause other problems when creating Dstream or
> creating the batch job.
> Even though the down of kafka broker is not the bug of spark streaming, we
> can avoid this failure in spark streaming. Especially for the reasons that
> kakfa cluster is not always stable in the real production.
> Does anyoner think we should add some retry logic when kakfa broker is down?
> I have implement this function in spark 1.6.3 and spark 2.1.0, and test them.
> If we implement this function, it will reduce the failure number of
> kafka-streaming which may help streaming users.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]