Re: Spark 1.0.0 rc3

2014-05-03 Thread Nan Zhu
SPARK_HADOOP_VERSION=2.3.0 sbt/sbt assembly and copy the generated jar to lib/ directory of my application, it seems that sbt cannot find the dependencies in the jar? but everything works with the pre-built jar files downloaded from the link provided by Patrick Best, -- Nan Zhu On

Apache Spark running out of the spark shell

2014-05-03 Thread Ajay Nair
Hi, I have written a code that works just about fine in the spark shell on EC2. The ec2 script helped me configure my master and worker nodes. Now I want to run the scala-spark code out side the interactive shell. How do I go about doing it. I was referring to the instructions mentioned here:

Re: Apache Spark running out of the spark shell

2014-05-03 Thread Sandy Ryza
Hi AJ, You might find this helpful - http://blog.cloudera.com/blog/2014/04/how-to-run-a-simple-apache-spark-app-in-cdh-5/ -Sandy On Sat, May 3, 2014 at 8:42 AM, Ajay Nair prodig...@gmail.com wrote: Hi, I have written a code that works just about fine in the spark shell on EC2. The ec2

Re: Apache Spark running out of the spark shell

2014-05-03 Thread Ajay Nair
Thank you for the reply. Have you posted a link from where I follow the steps ? -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/Apache-Spark-running-out-of-the-spark-shell-tp6459p6462.html Sent from the Apache Spark Developers List mailing list

Re: Apache Spark running out of the spark shell

2014-05-03 Thread Nicolas Garneau
Sorry, the link went wrong. I meant here: https://github.com/ngarneau/spark-standalone Le 2014-05-03 à 13:23, Nicolas Garneau ngarn...@ngarneau.com a écrit : Hey AJ, I created a little sample app using the spark's quick start. Have a look here. Assuming you used scala, using sbt is good

Re: Apache Spark running out of the spark shell

2014-05-03 Thread Ajay Nair
Thank you. Let me try this quickly ! -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/Apache-Spark-running-out-of-the-spark-shell-tp6459p6463.html Sent from the Apache Spark Developers List mailing list archive at Nabble.com.

Re: Apache Spark running out of the spark shell

2014-05-03 Thread Ajay Nair
Quick question, where should I place your folder. Inside the spark directory. My Spark directory is in /root/spark So currently I tried pulling your github code in /root/spark/spark-examples and modified my home spark directory in the scala code. I copied the sbt folder within the spark-examples

Re: Mailing list

2014-05-03 Thread Matei Zaharia
Hi Nicolas, Good catches on these things. Your website seems a little bit incomplete. I have found this page [1] with list the two main mailing lists, users and dev. But I see a reference to a mailing list about issues which tracks the sparks issues when it was hosted at Atlassian. I