GitHub user gaborgsomogyi opened a pull request:

    https://github.com/apache/spark/pull/23156

    [SPARK-24063][SS] Add maximum epoch queue threshold for ContinuousExecution

    ## What changes were proposed in this pull request?
    
    Continuous processing is waiting on epochs which are not yet complete (for 
example one partition is not making progress) and stores pending items in 
queues. These queues are unbounded and can consume up all the memory easily. In 
this PR I've added `spark.sql.streaming.continuous.epochBacklogQueueSize` 
configuration possibility to make them bounded. If the related threshold 
reached then the query will stop with `IllegalStateException`.
    
    ## How was this patch tested?
    
    Existing + additional unit tests.


You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/gaborgsomogyi/spark SPARK-24063

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/23156.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #23156
    
----
commit 72733c5068d85e70e0a65b05a593c82120277622
Author: Gabor Somogyi <gabor.g.somogyi@...>
Date:   2018-11-22T20:45:37Z

    [SPARK-24063][SS] Add maximum epoch queue threshold for ContinuousExecution.

----


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to