Hi all,

We are developing an application which uses Spark & Hive to do static and
ad-hoc reporting. For these static reports, they take a number of parameters
and then run over a data set. We would like to make it easier to test
performance of these reports on a cluster.

If we have a test cluster running with a sufficient sample data set which
developers can share. To speed up development time, what is the best way to
deploy a Spark application to a Spark cluster (in standalone) via an IDE?

I'm thinking we would create an SBT task which would run the spark submit
script. Is there a better way?

Eventually this will feed into some automated performance testing which we
plan to run as a twice daily Jenkins job. If its an SBT deploy task, it
makes it easy to call in Jenkins. Is there a better way to do this?

Posted on StackOverflow as well;
http://stackoverflow.com/questions/25048784/spark-automated-deployment-performance-testing
 

Any advice/experience appreciated!

Cheers!



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Deployment-Patterns-Automated-Deployment-Performance-Testing-tp11000.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to