we were thinking along the same line, that is to fix the number of streams
and change the input and output channels dynamically.

But could not make it work (seems that the receiver is not allowing any
change in the config after it started).

thanks,

On Wed, Jan 21, 2015 at 10:49 AM, Gerard Maas <gerard.m...@gmail.com> wrote:

> One possible workaround could be to orchestrate launch/stopping of
> Streaming jobs on demand as long as the number of jobs/streams stay
> within the boundaries of the resources (cores) you've available.
> e.g. if you're using Mesos, Marathon offers a REST interface to manage job
> lifecycle. You will still need to solve the dynamic configuration through
> some alternative channel.
>
> On Wed, Jan 21, 2015 at 11:30 AM, Tamas Jambor <jambo...@gmail.com> wrote:
>
>> thanks for the replies.
>>
>> is this something we can get around? Tried to hack into the code without
>> much success.
>>
>> On Wed, Jan 21, 2015 at 3:15 AM, Shao, Saisai <saisai.s...@intel.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I don't think current Spark Streaming support this feature, all the
>>> DStream lineage is fixed after the context is started.
>>>
>>> Also stopping a stream is not supported, instead currently we need to
>>> stop the whole streaming context to meet what you want.
>>>
>>> Thanks
>>> Saisai
>>>
>>> -----Original Message-----
>>> From: jamborta [mailto:jambo...@gmail.com]
>>> Sent: Wednesday, January 21, 2015 3:09 AM
>>> To: user@spark.apache.org
>>> Subject: dynamically change receiver for a spark stream
>>>
>>> Hi all,
>>>
>>> we have been trying to setup a stream using a custom receiver that would
>>> pick up data from sql databases. we'd like to keep that stream context
>>> running and dynamically change the streams on demand, adding and removing
>>> streams based on demand. alternativel, if a stream is fixed, is it possible
>>> to stop a stream, change to config and start again?
>>>
>>> thanks,
>>>
>>>
>>>
>>> --
>>> View this message in context:
>>> http://apache-spark-user-list.1001560.n3.nabble.com/dynamically-change-receiver-for-a-spark-stream-tp21268.html
>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For
>>> additional commands, e-mail: user-h...@spark.apache.org
>>>
>>>
>>
>

Reply via email to