If you're really doing a daily batch job, have you considered just using KafkaUtils.createRDD rather than a streaming job?
On Fri, Dec 18, 2015 at 5:04 AM, kali.tumm...@gmail.com < kali.tumm...@gmail.com> wrote: > Hi All, > > Imagine I have a Production spark streaming kafka (direct connection) > subscriber and publisher jobs running which publish and subscriber > (receive) > data from a kafka topic and I save one day's worth of data using > dstream.slice to Cassandra daily table (so I create daily table before > running spark streaming job). > > My question if all the above code runs in some scheduler like autosys how > should I say to spark publisher to stop publishing as it is End of day and > to spark subscriber to stop receiving to stop receiving without killing the > jobs ? if I kill my autosys scheduler turns red saying the job had failed > etc ... > Is there a way to stop both subscriber and publisher with out killing or > terminating the code. > > Thanks > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/how-to-turn-off-spark-streaming-gracefully-tp25734.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >