When running in standalone cluster mode, how can I verify that more than just 
one worker is being utilized? I can see that multiple workers are being started 
up, from the files in $SCALA_HOME/logs, but I don't see any difference in 
execution time when I specify 1 worker versus 4, which surprises me.

Here's my SparkConf as defined in my application:

val conf = new SparkConf().setAppName("My 
App").setMaster("local").setSparkHome("/home/me/spark-1.3.0-bin-hadoop2.4").setJars(List("target/scala-2.10/my-app-project_2.10-1.0.jar"))

And here's how I'm launching the master and slaves and submitting the job:

start-all.sh
spark-submit --class MyApp --master spark://localhost:7077 
target/scala-2.10/my-app-project_2.10-1.0.jar
stop-all.sh

In particular, I wonder if there's a problem because I specify "local" as the 
master inside of my application; perhaps that keeps the driver from sending any 
work to workers, even if they're registered? (Looking at the master and worker 
logs, the workers and the master do find each other: I get messages saying 
"Registering worker" and "Successfully registered with master", etc.)

Cheers,
Michael

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to