GitHub user sirishaSindri opened a pull request:
https://github.com/apache/spark/pull/20836
SPARK-23685 : Fix for the Spark Structured Streaming Kafka 0.10 Consuâ¦
â¦mer Can't Handle Non-consecutive Offsets
## What changes were proposed in this pull request?
In the fetchData , Instead of throwing an exception on failOnDataLoss,
I am saying return the record
if its offset falls in the user requested offset range
## How this patch was tested :
manually tested, added a unit test and ran in a real deployment
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/sirishaSindri/spark SPARK-23685
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/20836.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #20836
----
commit 5ccfed840f9cf9cd1c28a309b934e1285332d04d
Author: z001k5c <sirisha.sindiri@...>
Date: 2018-03-15T15:53:14Z
SPARK-23685 : Fix for the Spark Structured Streaming Kafka 0.10 Consumer
Can't Handle Non-consecutive Offsets
commit 7e08cd9062683c062b7b0408ffe40ff726249909
Author: z001k5c <sirisha.sindiri@...>
Date: 2018-03-15T17:06:06Z
SPARK-23685 : Fix for the Spark Structured Streaming Kafka 0.10 Consumer
Can't Handle Non-consecutive Offsets
----
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]