GitHub user viirya opened a pull request:

    https://github.com/apache/spark/pull/9665

    [SPARK-11698][Streaming] Add option to ignore kafka messages that are out 
of limit rate

    JIRA: https://issues.apache.org/jira/browse/SPARK-11698
    
    With spark.streaming.kafka.maxRatePerPartition, we can control the max rate 
limit. However, we can not ignore these messages out of limit. These messages 
will be consumed in next iteration. We have a use case that we need to ignore 
these messages and process latest messages in next iteration.
    
    In other words, we simply want to consume part of messages in each 
iteration and ignore remaining messages that are not consumed.
    
    We add an option for this purpose.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/viirya/spark-1 kafka-skip-records

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/9665.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #9665
    
----
commit d26e31b5591bf546f7c595ef81bc4027a726046e
Author: Liang-Chi Hsieh <[email protected]>
Date:   2015-11-12T16:04:55Z

    Add option to skip remaining records that are out of limit rate.

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to