You may consider writing back to Kafka from main stream and then have
downstream consumers.
This will keep things modular and independent.
On 15 Sep 2016 23:29, "Udbhav Agarwal" <udbhav.agar...@syncoms.com> wrote:

> Thank you Ayan for a reply.
>
> Source is kafka but I am reading from this source in my main stream. I
> will perform some operations here. Then I want to send the output of these
> operation to 4 parallel tasks. For these 4 parallel tasks I want 4 new
> streams. Is such an implementation possible here ?
>
>
>
> Thanks,
>
> Udbhav
>
> *From:* ayan guha [mailto:guha.a...@gmail.com]
> *Sent:* Thursday, September 15, 2016 6:43 PM
> *To:* Udbhav Agarwal <udbhav.agar...@syncoms.com>
> *Cc:* user <user@spark.apache.org>
> *Subject:* Re: Spark processing Multiple Streams from a single stream
>
>
>
> Depending on source. For example, if source is Kafka then you can write 4
> streaming consumers.
>
> On 15 Sep 2016 20:11, "Udbhav Agarwal" <udbhav.agar...@syncoms.com> wrote:
>
> Hi All,
>
> I have a scenario where I want to process a message in various ways in
> parallel. For instance a message is coming inside spark stream(DStream) and
> I want to send this message to 4 different tasks in parallel. I want these
> 4 different tasks to be separate streams in the original spark stream and
> are always active and waiting for input. Can I implement such a process
> with spark streaming ? How ?
>
> Thanks in advance.
>
>
>
> *Thanks,*
>
> *Udbhav Agarwal*
>
>
>
>
>
>

Reply via email to