Hi Gerard, thanks, that makes sense. I'll try that out.
Tamas On Wed, Jan 21, 2015 at 11:14 AM, Gerard Maas <gerard.m...@gmail.com> wrote: > Hi Tamas, > > I meant not changing the receivers, but starting/stopping the Streaming > jobs. So you would have a 'small' Streaming job for a subset of streams > that you'd configure->start->stop on demand. > I haven't tried myself yet, but I think it should also be possible to > create a Streaming Job from the Spark Job Server ( > https://github.com/spark-jobserver/spark-jobserver). Then you would have > a REST interface that even gives you the possibility of passing a > configuration. > > -kr, Gerard. > > On Wed, Jan 21, 2015 at 11:54 AM, Tamas Jambor <jambo...@gmail.com> wrote: > >> we were thinking along the same line, that is to fix the number of >> streams and change the input and output channels dynamically. >> >> But could not make it work (seems that the receiver is not allowing any >> change in the config after it started). >> >> thanks, >> >> On Wed, Jan 21, 2015 at 10:49 AM, Gerard Maas <gerard.m...@gmail.com> >> wrote: >> >>> One possible workaround could be to orchestrate launch/stopping of >>> Streaming jobs on demand as long as the number of jobs/streams stay >>> within the boundaries of the resources (cores) you've available. >>> e.g. if you're using Mesos, Marathon offers a REST interface to manage >>> job lifecycle. You will still need to solve the dynamic configuration >>> through some alternative channel. >>> >>> On Wed, Jan 21, 2015 at 11:30 AM, Tamas Jambor <jambo...@gmail.com> >>> wrote: >>> >>>> thanks for the replies. >>>> >>>> is this something we can get around? Tried to hack into the code >>>> without much success. >>>> >>>> On Wed, Jan 21, 2015 at 3:15 AM, Shao, Saisai <saisai.s...@intel.com> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> I don't think current Spark Streaming support this feature, all the >>>>> DStream lineage is fixed after the context is started. >>>>> >>>>> Also stopping a stream is not supported, instead currently we need to >>>>> stop the whole streaming context to meet what you want. >>>>> >>>>> Thanks >>>>> Saisai >>>>> >>>>> -----Original Message----- >>>>> From: jamborta [mailto:jambo...@gmail.com] >>>>> Sent: Wednesday, January 21, 2015 3:09 AM >>>>> To: user@spark.apache.org >>>>> Subject: dynamically change receiver for a spark stream >>>>> >>>>> Hi all, >>>>> >>>>> we have been trying to setup a stream using a custom receiver that >>>>> would pick up data from sql databases. we'd like to keep that stream >>>>> context running and dynamically change the streams on demand, adding and >>>>> removing streams based on demand. alternativel, if a stream is fixed, is >>>>> it >>>>> possible to stop a stream, change to config and start again? >>>>> >>>>> thanks, >>>>> >>>>> >>>>> >>>>> -- >>>>> View this message in context: >>>>> http://apache-spark-user-list.1001560.n3.nabble.com/dynamically-change-receiver-for-a-spark-stream-tp21268.html >>>>> Sent from the Apache Spark User List mailing list archive at >>>>> Nabble.com. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For >>>>> additional commands, e-mail: user-h...@spark.apache.org >>>>> >>>>> >>>> >>> >> >