Thanks Stephen. That did the job for me. After adding JAVA_HOME in hadoop-layout.sh, I was able to run the spark-job successfully. @mgummelt - I did not set the executor_environment_variables. However, I am now able to see JAVA_HOME when I print out the 'env' in the driver.
On Fri, Aug 5, 2016 at 1:53 PM, [email protected] < [email protected]> wrote: > What is your --executor-environment-variables set to? > http://mesos.apache.org/documentation/latest/configuration/ > > Can you print out your `env` in the driver to verify it has the expected > JAVA_HOME > > On 2016-08-04 12:28 (-0700), max square <[email protected]> wrote: > > Hey guys , > > I was trying out spark 2.0 examples to run on mesos+hadoop cluster but it > > keep failing with the following error message:- > > > > I0803 19:46:53.848696 12494 fetcher.cpp:498] Fetcher Info: > > > {"cache_directory":"\/tmp\/mesos\/fetch\/slaves\/ > 587226cc-bece-422a-bb93-e3ef49075642-S1\/root","items" > :[{"action":"BYPASS_CACHE","uri":{"extract":true,"value":" > hdfs:\/\/testcluster\/spark-examples_2.11-2.0.0.jar"}},{" > action":"BYPASS_CACHE","uri":{"extract":true,"value":"hdfs:\ > /\/testcluster\/spark-2.0.0-bin-hdfs-2.6.0-cdh5.7.1.tgz"}} > ],"sandbox_directory":"\/vol\/mesos\/data\/slaves\/587226cc- > bece-422a-bb93-e3ef49075642-S1\/frameworks\/587226cc-bece- > 422a-bb93-e3ef49075642-0017\/executors\/driver-20160803194649-0001\/runs\/ > b1e9a92e-f004-4cdc-b936-52b32593d39f","user":"root"} > > > > I0803 19:46:53.850719 12494 fetcher.cpp:409] Fetching URI > > > 'hdfs://testcluster/spark-examples_2.11-2.0.0.jar' > > > > I0803 19:46:53.850731 12494 fetcher.cpp:250] Fetching directly into the > > > sandbox directory > > > > I0803 19:46:53.850746 12494 fetcher.cpp:187] Fetching URI > > > 'hdfs://testcluster/spark-examples_2.11-2.0.0.jar' > > > E0803 19:46:53.860776 12494 shell.hpp:106] Command > > > '/usr/lib/hadoop/bin/hadoop version 2>&1' failed; this is the output: > > > Error: JAVA_HOME is not set and could not be found. > > > Failed to fetch 'hdfs://testcluster/spark-examples_2.11-2.0.0.jar': > Failed > > > to create HDFS client: Failed to execute '/usr/lib/hadoop/bin/hadoop > > > version 2>&1'; the command was either not found or exited with a > non-zero > > > exit status: 1 > > > Failed to synchronize with agent (it's probably exited) > > > > > > To start out, I tried out the hadoop command which was giving the error > on > > the agents and was able to replicate the error. So basically, running > "sudo > > -u root /usr/lib/hadoop/bin/hadoop version 2>&1" gave me the same > JAVA_HOME > > not set error. After I fixed that and restarted the agents, running the > > spark example still gave me the same error. > > > > I ran the same examples on mesos 0.28.2, and it ran fine. > > > > Any help regarding this would be appreciated. > > > > *Additional Info :-* > > mesos version - 1.0.0 > > hadoop version - 2.6.0-cdh5.7.2 > > spark version - 2.0.0 > > > > Command used to run spark example - ./bin/spark-submit --class > > org.apache.spark.examples.SparkPi --master mesos://<master>:7077 > > --deploy-mode cluster --executor-memory 2G --total-executor-cores 4 > > hdfs://testcluster/spark-examples_2.11-2.0.0.jar 100 > > >

