Hi All, Thank you for the response.
As per https://docs.cloud.databricks.com/docs/latest/databricks_guide/index.html#07%20Spark%20Streaming/15%20Streaming%20FAQs.html There can be only one streaming context in a cluster which implies only one streaming job. So, I am still confused. Anyone having more than 1 spark streaming app in a cluster running at the same time, please share your experience. Thanks On Wed, Dec 14, 2016 at 6:54 PM, Akhilesh Pathodia < pathodia.akhil...@gmail.com> wrote: > If you have enough cores/resources, run them separately depending on your > use case. > > > On Thursday 15 December 2016, Divya Gehlot <divya.htco...@gmail.com> > wrote: > >> It depends on the use case ... >> Spark always depends on the resource availability . >> As long as you have resource to acoomodate ,can run as many spark/spark >> streaming application. >> >> >> Thanks, >> Divya >> >> On 15 December 2016 at 08:42, shyla deshpande <deshpandesh...@gmail.com> >> wrote: >> >>> How many Spark streaming applications can be run at a time on a Spark >>> cluster? >>> >>> Is it better to have 1 spark streaming application to consume all the >>> Kafka topics or have multiple streaming applications when possible to keep >>> it simple? >>> >>> Thanks >>> >>> >>