Sorry, the link went wrong. I meant here: https://github.com/ngarneau/spark-standalone
Le 2014-05-03 à 13:23, Nicolas Garneau <ngarn...@ngarneau.com> a écrit : > Hey AJ, > > I created a little sample app using the spark's quick start. > Have a look here. > Assuming you used scala, using sbt is good for running your application in > standalone mode. > The configuration file which is "simple.sbt" in my repo, holds all the > dependencies needed to build your app. > > Hope this helps! > > Le 2014-05-03 à 11:42, Ajay Nair <prodig...@gmail.com> a écrit : > >> Hi, >> >> I have written a code that works just about fine in the spark shell on EC2. >> The ec2 script helped me configure my master and worker nodes. Now I want to >> run the scala-spark code out side the interactive shell. How do I go about >> doing it. >> >> I was referring to the instructions mentioned here: >> https://spark.apache.org/docs/0.9.1/quick-start.html >> >> But this is confusing because it mentions about a simple project jar file >> which I am not sure how to generate. I only have the file that runs directly >> on my spark shell. Any easy intruction to get this quickly running as a job? >> >> Thanks >> AJ >> >> >> >> -- >> View this message in context: >> http://apache-spark-developers-list.1001551.n3.nabble.com/Apache-Spark-running-out-of-the-spark-shell-tp6459.html >> Sent from the Apache Spark Developers List mailing list archive at >> Nabble.com. >> > > Nicolas Garneau > ngarn...@ngarneau.com > Nicolas Garneau 418.569.3097 ngarn...@ngarneau.com