http://spark.apache.org/docs/latest/submitting-applications.html
look at cluster mode, supervise On Fri, Jul 22, 2016 at 8:46 AM, Sivakumaran S <siva.kuma...@me.com> wrote: > Hello, > > I have a spark streaming process on a cluster ingesting a realtime data > stream from Kafka. The aggregated, processed output is written to Cassandra > and also used for dashboard display. > > My question is - If the node running the driver program fails, I am guessing > that the entire process fails and has to be restarted. Is there any way to > obviate this? Is my understanding correct that the spark-submit in its > current form is a Single Point of Vulnerability, much akin to the NameNode in > HDFS? > > regards > > Sivakumaran S > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org