Hi,
each node has 4Gig of memory. After total reboot and re-run of SparkPi  
resource manager shows no running containers and 1 pending container. 

-jan

On 20 May 2014, at 10:24, <sandy.r...@cloudera.com> <sandy.r...@cloudera.com> 
wrote:

> Hi Jan,
> 
> How much memory capacity is configured for each node?
> 
> If you go to the ResourceManager web UI, does it indicate any containers are 
> running?
> 
> -Sandy
> 
>> On May 19, 2014, at 11:43 PM, Jan Holmberg <jan.holmb...@perigeum.fi> wrote:
>> 
>> Hi,
>> I’m new to Spark and trying to test first Spark prog. I’m running SparkPi 
>> successfully in yarn-client -mode but when running the same in yarn-mode, 
>> app gets stuck to ACCEPTED phase. I’ve tried hours to hunt down the reason 
>> but the outcome is always the same. Any hints what to look for next? 
>> 
>> cheers,
>> -jan
>> 
>> 
>> vagrant@vm-cluster-node1:~$ ./run_pi.sh
>> 14/05/20 06:24:04 INFO RMProxy: Connecting to ResourceManager at 
>> vm-cluster-node2/10.211.55.101:8032
>> 14/05/20 06:24:05 INFO Client: Got Cluster metric info from 
>> ApplicationsManager (ASM), number of NodeManagers: 2
>> 14/05/20 06:24:05 INFO Client: Queue info ... queueName: root.default, 
>> queueCurrentCapacity: 0.0, queueMaxCapacity: -1.0,
>>     queueApplicationCount = 0, queueChildQueueCount = 0
>> 14/05/20 06:24:05 INFO Client: Max mem capabililty of a single resource in 
>> this cluster 2048
>> 14/05/20 06:24:05 INFO Client: Preparing Local resources
>> 14/05/20 06:24:05 INFO Client: Uploading 
>> file:/opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/spark/assembly/lib/spark-assembly_2.10-0.9.0-cdh5.0.0-hadoop2.3.0-cdh5.0.0.jar
>>  to 
>> hdfs://vm-cluster-node2:8020/user/vagrant/.sparkStaging/application_1400563733088_0012/spark-assembly_2.10-0.9.0-cdh5.0.0-hadoop2.3.0-cdh5.0.0.jar
>> 14/05/20 06:24:07 INFO Client: Setting up the launch environment
>> 14/05/20 06:24:07 INFO Client: Setting up container launch context
>> 14/05/20 06:24:07 INFO Client: Command for starting the Spark 
>> ApplicationMaster: java -server -Xmx1024m -Djava.io.tmpdir=$PWD/tmp 
>> org.apache.spark.deploy.yarn.ApplicationMaster --class 
>> org.apache.spark.examples.SparkPi --jar 
>> /opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/spark/assembly/lib/spark-assembly_2.10-0.9.0-cdh5.0.0-hadoop2.3.0-cdh5.0.0.jar
>>  --args  'yarn-standalone'  --args  '10'  --worker-memory 500 --worker-cores 
>> 1 --num-workers 1 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr
>> 14/05/20 06:24:07 INFO Client: Submitting application to ASM
>> 14/05/20 06:24:07 INFO YarnClientImpl: Submitted application 
>> application_1400563733088_0012
>> 14/05/20 06:24:08 INFO Client: Application report from ASM: <THIS PART GET 
>> REPEATING FOREVER>
>>    application identifier: application_1400563733088_0012
>>    appId: 12
>>    clientToAMToken: null
>>    appDiagnostics:
>>    appMasterHost: N/A
>>    appQueue: root.vagrant
>>    appMasterRpcPort: -1
>>    appStartTime: 1400567047343
>>    yarnAppState: ACCEPTED
>>    distributedFinalState: UNDEFINED
>>    appTrackingUrl: 
>> http://vm-cluster-node2:8088/proxy/application_1400563733088_0012/
>>    appUser: vagrant
>> 
>> 
>> Log files give me no additional help. Latest log entry just acknowledges the 
>> status change:
>> 
>> hadoop-yarn/hadoop-cmf-yarn-RESOURCEMANAGER-vm-cluster-node2.log.out:2014-05-20
>>  06:24:07,347 INFO 
>> org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: 
>> application_1400563733088_0012 State change from SUBMITTED to ACCEPTED
>> 
>> 
>> I’m running the example in local test environment with three virtual nodes 
>> in Cloudera (CDH5).
>> 
>> Below is the run_pi.sh : 
>> 
>> #!/bin/bash
>> 
>> export SPARK_HOME=/opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/spark
>> export STANDALONE_SPARK_MASTER_HOST=vm-cluster-node2
>> export SPARK_MASTER_PORT=7077
>> export 
>> DEFAULT_HADOOP_HOME=/opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hadoop
>> 
>> export SPARK_JAR_HDFS_PATH=/user/spark/share/lib/spark-assembly.jar
>> 
>> export SPARK_LAUNCH_WITH_SCALA=0
>> export SPARK_LIBRARY_PATH=${SPARK_HOME}/lib
>> export SCALA_LIBRARY_PATH=${SPARK_HOME}/lib
>> export SPARK_MASTER_IP=$STANDALONE_SPARK_MASTER_HOST
>> 
>> export HADOOP_HOME=${HADOOP_HOME:-$DEFAULT_HADOOP_HOME}
>> 
>> if [ -n "$HADOOP_HOME" ]; then
>> export SPARK_LIBRARY_PATH=$SPARK_LIBRARY_PATH:${HADOOP_HOME}/lib/native
>> fi
>> export 
>> SPARK_JAR=hdfs://vm-cluster-node2:8020/user/spark/share/lib/spark-assembly.jar
>> 
>> APP_JAR=/opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/spark/assembly/lib/spark-assembly_2.10-0.9.0-cdh5.0.0-hadoop2.3.0-cdh5.0.0.jar
>> 
>> $SPARK_HOME/bin/spark-class org.apache.spark.deploy.yarn.Client \
>> --jar $APP_JAR \
>> --class org.apache.spark.examples.SparkPi \
>> --args yarn-standalone \
>> --args 10 \
>> --num-workers 1 \
>> --master-memory 1g \
>> --worker-memory 500m \
>> --worker-cores 1
>> 
>> 
>> 

Reply via email to