After the call to ssc.stop, you will probably see output printed into
Zeppelin; you should see output every time a window closes.

*FA*

*FA *
http://queirozf.com
“Every time you stay out late; every time you sleep in; every time you miss
a workout; every time you don’t give 100% – You make it that much easier
for me to beat you.” - Unknown author

On 15 February 2016 at 23:31, Felipe Almeida <falmeida1...@gmail.com> wrote:

> Yes there is. I do it every day. It's very useful to try things out to see
> if things are behaving according to plan.
>
> Create the spark streaming context from the sc SparkContext (I'm using
> @transient for everything (because I run stuff outside of any enclosing
> scope, like an object) stuff gets "sucked up" into the context and has to
> be serialized, which causes lots of errors)
>
> @transient val ssc = new StreamingContext(sc,windowDuration)
>
>
> Here is an example of stuff you can do (I'm consuming Kinesis streams, you
> might use something else)
>
> @transient val streams = (0 until 1).map { i =>
>   KinesisUtils.createStream(ssc, appName, streamName, endpointUrl, regionName,
>     InitialPositionInStream.LATEST, windowDuration, StorageLevel.MEMORY_ONLY)
> }
> @transient val events = ssc
> .union(streams)
> .map(byteArray => new String(byteArray))
> .map(str => Funcs.parseLogEvent(str))
> .filter(tryLogEvent => tryLogEvent.isSuccess)
> .map(tryLogEvent => tryLogEvent.get)
> .filter(logEvent => logEvent.DataType == "Event" && logEvent.RequestType != 
> "UIInputError" && logEvent.RequestType != "UIEmptyInput")
>
>
> You need an action like print() otherwise nothing happens, you probably
> know that
>
> events.print(10)
>
>
> Now you start the context:
>
> ssc.start()
>
>
> Now you need to wait for as long as it takes for a window to "close", and
> then you stop the streaming context, *but not the underlying SparkContext*,
> otherwise you need to restart the interpreter:
>
> ssc.stop(stopSparkContext=false, stopGracefully=true)
>
>
> Everything should work I guess. Let me know if you run into other problems.
>
> *FA*
>
>
> *FA *
> http://queirozf.com
> “Every time you stay out late; every time you sleep in; every time you
> miss a workout; every time you don’t give 100% – You make it that much
> easier for me to beat you.” - Unknown author
>
> On 15 February 2016 at 23:15, Michael Gummelt <mgumm...@mesosphere.io>
> wrote:
>
>> Hi,
>>
>> I've seen this JIRA regarding Spark Streaming:
>> https://issues.apache.org/jira/browse/ZEPPELIN-274, but I have a more
>> basic question.  Is there a way to iterate on a Spark Streaming job in
>> Zeppelin without restarting the interpreter?  Specifically, I'd like to
>> start a streaming job, see some results, rewrite the job, then start it
>> again.  Spark has some limitations that make this difficult:
>> http://spark.apache.org/docs/latest/streaming-programming-guide.html#points-to-remember
>>
>> But I'm wondering if there's a workaround.  Thanks.
>>
>> --
>> Michael Gummelt
>> Software Engineer
>> Mesosphere
>>
>
>

Reply via email to