Thanks for information this fixed the issue. Issue was in spark-master
memory when I specify manually 1G for master. it start working

On 30 July 2015 at 14:26, Shao, Saisai <saisai.s...@intel.com> wrote:

>  You’d better also check the log of nodemanager, sometimes because your
> memory usage exceeds the limit of Yarn container’s configuration.
>
>
>
> I’ve met similar problem before, here is the warning log in nodemanager:
>
>
>
> 2015-07-07 17:06:07,141 WARN
> org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
> Container [pid=17385,containerID=container_1436259427993_0001_02_000001] is
> running beyond virtual memory limits. Current usage: 318.1 MB of 1 GB
> physical memory used; 2.2 GB of 2.1 GB virtual memory used. Killing
> container.
>
>
>
> The default pmem-vmem ratio is 2.1, but seems executor requires more vmem
> when started, so nodemanager will kill it. If you met similar problem, you
> could increase this configuration “yarn.nodemanager.vmem-pmem-ratio”.
>
>
>
> Thanks
>
> Jerry
>
>
>
> *From:* Jeff Zhang [mailto:zjf...@gmail.com]
> *Sent:* Thursday, July 30, 2015 4:36 PM
> *To:* Jeetendra Gangele
> *Cc:* user
> *Subject:* Re: Spark on YARN
>
>
>
> >> 15/07/30 12:13:35 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15:
> SIGTERM
>
>
>
> AM is killed somehow, may due to preemption. Does it always happen ?
> Resource manager log would be helpful.
>
>
>
>
>
>
>
> On Thu, Jul 30, 2015 at 4:17 PM, Jeetendra Gangele <gangele...@gmail.com>
> wrote:
>
>  I can't see the application logs here. All the logs are going into
> stderr. can anybody help here?
>
>
>
> On 30 July 2015 at 12:21, Jeetendra Gangele <gangele...@gmail.com> wrote:
>
>  I am running below command this is default spark PI program but this is
> not running all the log are going in stderr but at the terminal job is
> succeeding .I guess there are con issue job it not at all launching
>
>
>
> /bin/spark-submit --class org.apache.spark.examples.SparkPi --master
> yarn-cluster lib/spark-examples-1.4.1-hadoop2.6.0.jar 10
>
>
>
>
>
> Complete log
>
>
>
> SLF4J: Class path contains multiple SLF4J bindings.
>
> SLF4J: Found binding in 
> [jar:file:/home/hadoop/tmp/nm-local-dir/usercache/hadoop/filecache/23/spark-assembly-1.4.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>
> SLF4J: Found binding in 
> [jar:file:/opt/hadoop-2.7.0/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
>
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>
> 15/07/30 12:13:31 INFO yarn.ApplicationMaster: Registered signal handlers for 
> [TERM, HUP, INT]
>
> 15/07/30 12:13:32 INFO yarn.ApplicationMaster: ApplicationAttemptId: 
> appattempt_1438090734187_0010_000001
>
> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing view acls to: hadoop
>
> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing modify acls to: hadoop
>
> 15/07/30 12:13:33 INFO spark.SecurityManager: SecurityManager: authentication 
> disabled; ui acls disabled; users with view permissions: Set(hadoop); users 
> with modify permissions: Set(hadoop)
>
> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Starting the user application 
> in a separate Thread
>
> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Waiting for spark context 
> initialization
>
> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Waiting for spark context 
> initialization ...
>
> 15/07/30 12:13:33 INFO spark.SparkContext: Running Spark version 1.4.1
>
> 15/07/30 12:13:33 WARN spark.SparkConf:
>
> SPARK_JAVA_OPTS was detected (set to '-Dspark.driver.port=53411').
>
> This is deprecated in Spark 1.0+.
>
>
>
> Please instead use:
>
>  - ./spark-submit with conf/spark-defaults.conf to set defaults for an 
> application
>
>  - ./spark-submit with --driver-java-options to set -X options for a driver
>
>  - spark.executor.extraJavaOptions to set -X options for executors
>
>  - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master 
> or worker)
>
>
>
> 15/07/30 12:13:33 WARN spark.SparkConf: Setting 
> 'spark.executor.extraJavaOptions' to '-Dspark.driver.port=53411' as a 
> work-around.
>
> 15/07/30 12:13:33 WARN spark.SparkConf: Setting 
> 'spark.driver.extraJavaOptions' to '-Dspark.driver.port=53411' as a 
> work-around.
>
> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing view acls to: hadoop
>
> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing modify acls to: hadoop
>
> 15/07/30 12:13:33 INFO spark.SecurityManager: SecurityManager: authentication 
> disabled; ui acls disabled; users with view permissions: Set(hadoop); users 
> with modify permissions: Set(hadoop)
>
> 15/07/30 12:13:33 INFO slf4j.Slf4jLogger: Slf4jLogger started
>
> 15/07/30 12:13:33 INFO Remoting: Starting remoting
>
> 15/07/30 12:13:34 INFO Remoting: Remoting started; listening on addresses 
> :[akka.tcp://sparkDriver@10.21.1.77:53411]
>
> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'sparkDriver' 
> on port 53411.
>
> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering MapOutputTracker
>
> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering BlockManagerMaster
>
> 15/07/30 12:13:34 INFO storage.DiskBlockManager: Created local directory at 
> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/blockmgr-2166bbd9-b1ed-41d1-bc95-92c6a7fbd36f
>
> 15/07/30 12:13:34 INFO storage.MemoryStore: MemoryStore started with capacity 
> 246.0 MB
>
> 15/07/30 12:13:34 INFO spark.HttpFileServer: HTTP File server directory is 
> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/httpd-d1232310-5aa1-44e7-a99a-cc2ae614f89c
>
> 15/07/30 12:13:34 INFO spark.HttpServer: Starting HTTP Server
>
> 15/07/30 12:13:34 INFO server.Server: jetty-8.y.z-SNAPSHOT
>
> 15/07/30 12:13:34 INFO server.AbstractConnector: Started 
> SocketConnector@0.0.0.0:52507
>
> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'HTTP file 
> server' on port 52507.
>
> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering OutputCommitCoordinator
>
> 15/07/30 12:13:34 INFO ui.JettyUtils: Adding filter: 
> org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
>
> 15/07/30 12:13:34 INFO server.Server: jetty-8.y.z-SNAPSHOT
>
> 15/07/30 12:13:34 INFO server.AbstractConnector: Started 
> SelectChannelConnector@0.0.0.0:59596
>
> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'SparkUI' on 
> port 59596.
>
> 15/07/30 12:13:34 INFO ui.SparkUI: Started SparkUI at http://10.21.1.77:59596
>
> 15/07/30 12:13:34 INFO cluster.YarnClusterScheduler: Created 
> YarnClusterScheduler
>
> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43354.
>
> 15/07/30 12:13:34 INFO netty.NettyBlockTransferService: Server created on 
> 43354
>
> 15/07/30 12:13:34 INFO storage.BlockManagerMaster: Trying to register 
> BlockManager
>
> 15/07/30 12:13:34 INFO storage.BlockManagerMasterEndpoint: Registering block 
> manager 10.21.1.77:43354 with 246.0 MB RAM, BlockManagerId(driver, 
> 10.21.1.77, 43354)
>
> 15/07/30 12:13:34 INFO storage.BlockManagerMaster: Registered BlockManager
>
> 15/07/30 12:13:34 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: 
> ApplicationMaster registered as 
> AkkaRpcEndpointRef(Actor[akka://sparkDriver/user/YarnAM#-603094240])
>
> 15/07/30 12:13:34 INFO client.RMProxy: Connecting to ResourceManager at 
> hadoop-1/10.21.1.77:8030
>
> 15/07/30 12:13:34 INFO yarn.YarnRMClient: Registering the ApplicationMaster
>
> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Will request 2 executor 
> containers, each with 1 cores and 1408 MB memory including 384 MB overhead
>
> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Container request (host: Any, 
> capability: <memory:1408, vCores:1>)
>
> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Container request (host: Any, 
> capability: <memory:1408, vCores:1>)
>
> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Started progress reporter 
> thread - sleep time : 5000
>
> 15/07/30 12:13:35 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: SIGTERM
>
> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Final app status: SUCCEEDED, 
> exitCode: 0, (reason: Shutdown hook called before final status was reported.)
>
> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Unregistering 
> ApplicationMaster with SUCCEEDED (diag message: Shutdown hook called before 
> final status was reported.)
>
> 15/07/30 12:13:35 INFO impl.AMRMClientImpl: Waiting for application to be 
> successfully unregistered.
>
> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Deleting staging directory 
> .sparkStaging/application_1438090734187_0010
>
> 15/07/30 12:13:35 INFO storage.DiskBlockManager: Shutdown hook called
>
> 15/07/30 12:13:35 INFO util.Utils: Shutdown hook called
>
> 15/07/30 12:13:35 INFO util.Utils: Deleting directory 
> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/userFiles-337c9be5-569f-43ff-ba1f-ec24daab9ea5
>
> 15/07/30 12:13:35 INFO util.Utils: Deleting directory 
> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/httpd-d1232310-5aa1-44e7-a99a-cc2ae614f89c
>
>
>
>
>
>
>
> --
>
> Best Regards
>
> Jeff Zhang
>

Reply via email to