Where it could start to make some sense is if you wanted a single application to be able to work with more than one Spark cluster -- but that's a pretty weird or unusual thing to do, and I'm pretty sure it wouldn't work correctly at present.
On Fri, Dec 4, 2015 at 11:10 AM, Michael Armbrust <mich...@databricks.com> wrote: > To be clear, I don't think there is ever a compelling reason to create > more than one SparkContext in a single application. The context is > threadsafe and can launch many jobs in parallel from multiple threads. > Even if there wasn't global state that made it unsafe to do so, creating > more than one context creates an artificial barrier that prevents sharing > of RDDs between the two. > > On Fri, Dec 4, 2015 at 10:47 AM, prateek arora <prateek.arora...@gmail.com > > wrote: > >> Thanks ... >> >> Is there any way my second application run in parallel and wait for >> fetching data from hbase or any other data storeage system ? >> >> Regards >> Prateek >> >> On Fri, Dec 4, 2015 at 10:24 AM, Ted Yu <yuzhih...@gmail.com> wrote: >> >>> How about using NoSQL data store such as HBase :-) >>> >>> On Fri, Dec 4, 2015 at 10:17 AM, prateek arora < >>> prateek.arora...@gmail.com> wrote: >>> >>>> Hi Ted >>>> Thanks for the information . >>>> is there any way that two different spark application share there data ? >>>> >>>> Regards >>>> Prateek >>>> >>>> On Fri, Dec 4, 2015 at 9:54 AM, Ted Yu <yuzhih...@gmail.com> wrote: >>>> >>>>> See Josh's response in this thread: >>>>> >>>>> >>>>> http://search-hadoop.com/m/q3RTt1z1hUw4TiG1&subj=Re+Question+about+yarn+cluster+mode+and+spark+driver+allowMultipleContexts >>>>> >>>>> Cheers >>>>> >>>>> On Fri, Dec 4, 2015 at 9:46 AM, prateek arora < >>>>> prateek.arora...@gmail.com> wrote: >>>>> >>>>>> Hi >>>>>> >>>>>> I want to create multiple sparkContext in my application. >>>>>> i read so many articles they suggest " usage of multiple contexts is >>>>>> discouraged, since SPARK-2243 is still not resolved." >>>>>> i want to know that Is spark 1.5.0 supported to create multiple >>>>>> contexts >>>>>> without error ? >>>>>> and if supported then are we need to set >>>>>> "spark.driver.allowMultipleContexts" configuration parameter ? >>>>>> >>>>>> Regards >>>>>> Prateek >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> View this message in context: >>>>>> http://apache-spark-user-list.1001560.n3.nabble.com/is-Multiple-Spark-Contexts-is-supported-in-spark-1-5-0-tp25568.html >>>>>> Sent from the Apache Spark User List mailing list archive at >>>>>> Nabble.com. >>>>>> >>>>>> --------------------------------------------------------------------- >>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>>> >>>>>> >>>>> >>>> >>> >> >