During the last few days I've been trying to deploy a Scala job to a standalone cluster (master + 4 workers) without much success, although it worked perfectly when launching it from the spark shell, that is, using the Scala REPL (pretty strange, this would mean my cluster config was actually correct).
In order to test it with a simpler example, I decided to deploy this example <https://spark.apache.org/docs/0.9.0/quick-start.html#a-standalone-app-in-scala> in standalone mode(master + 1 worker, same machine). Please have a look at this gist <https://gist.github.com/JordiAranda/4ee54f84dc92f02ecb8c> for the cluster setup. I can't get rid of the EOFException. So, I should definitely be missing something. Why it works when setting the master config property to "local[x]" or launching it from the REPL, and not when setting the master config property as an spark url? PS: Please, notice I am using the latest release (0.9.1) prebuilt for Hadoop 2 Thanks, -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/A-Standalone-App-in-Scala-Standalone-mode-issues-tp6493.html Sent from the Apache Spark User List mailing list archive at Nabble.com.