This is a known issue introduced in Spark 1.4.1 and 1.5.0 and will be fixed
in Spark 1.5.1. In the mean time, you could prototype in Spark 1.4.0 and
wait for Spark 1.5.1/1.4.2 to come out. You could also download the source
code and compile the Spark master branch.
https://issues.apache.org/jira/browse/SPARK-10071



On Tue, Sep 8, 2015 at 6:46 AM, Bryan Jeffrey <bryan.jeff...@gmail.com>
wrote:

> Hello. We're getting started with Spark Streaming. We're working to build
> some unit/acceptance testing around functions that consume DStreams. The
> current method for creating DStreams is to populate the data by creating an
> InputDStream:
>
> val input = Array(TestDataFactory.CreateEvent(123 notFoundData))
> val queue =
> scala.collection.mutable.Queue(ssc.sparkContext.parallelize(input))
> val events: InputDStream[MyEvent] = ssc.queueStream(queue)
>
> The 'events' InputDStream can then be fed into functions. However, the
> stream does not allow checkpointing. This means that we're unable to use
> this to feed methods/classes that execute stateful actions like
> 'updateStateByKey'.
>
> Does anyone have a simple, contained method to create DStreams that allow
> for checkpointing? I looked at the Spark unit test framework, but that
> seems to require access to a bunch of spark internals (requiring that
> you're within the spark package, etc.)
>
>

Reply via email to