It may be worthwhile to do architect the computation in a different way. dstream.foreachRDD { rdd => rdd.foreach { record => // do different things for each record based on filters } }
TD On Sun, Apr 12, 2015 at 7:52 PM, Jianshi Huang <jianshi.hu...@gmail.com> wrote: > Hi, > > I have a Kafka topic that contains dozens of different types of messages. > And for each one I'll need to create a DStream for it. > > Currently I have to filter the Kafka stream over and over, which is very > inefficient. > > So what's the best way to do dispatching in Spark Streaming? (one DStream > -> multiple DStreams) > > > Thanks, > -- > Jianshi Huang > > LinkedIn: jianshi > Twitter: @jshuang > Github & Blog: http://huangjs.github.com/ >