[jira] [Commented] (SPARK-23660) Yarn throws exception in cluster mode when the application is small
[ https://issues.apache.org/jira/browse/SPARK-23660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16396541#comment-16396541 ] Apache Spark commented on SPARK-23660: -- User 'gaborgsomogyi' has created a pull request for this issue: https://github.com/apache/spark/pull/20807 > Yarn throws exception in cluster mode when the application is small > --- > > Key: SPARK-23660 > URL: https://issues.apache.org/jira/browse/SPARK-23660 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.3.0 >Reporter: Gabor Somogyi >Priority: Minor > > Yarn throws the following exception in cluster mode when the application is > really small: > {code:java} > 18/03/07 23:34:22 WARN netty.NettyRpcEnv: Ignored failure: > java.util.concurrent.RejectedExecutionException: Task > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask@7c974942 > rejected from > java.util.concurrent.ScheduledThreadPoolExecutor@1eea9d2d[Terminated, pool > size = 0, active threads = 0, queued tasks = 0, completed tasks = 0] > 18/03/07 23:34:22 ERROR yarn.ApplicationMaster: Uncaught exception: > org.apache.spark.SparkException: Exception thrown in awaitResult: > at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:205) > at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75) > at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92) > at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76) > at > org.apache.spark.deploy.yarn.YarnAllocator.(YarnAllocator.scala:102) > at > org.apache.spark.deploy.yarn.YarnRMClient.register(YarnRMClient.scala:77) > at > org.apache.spark.deploy.yarn.ApplicationMaster.registerAM(ApplicationMaster.scala:450) > at > org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:493) > at > org.apache.spark.deploy.yarn.ApplicationMaster.org$apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:345) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply$mcV$sp(ApplicationMaster.scala:260) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anon$5.run(ApplicationMaster.scala:810) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1920) > at > org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:809) > at > org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:259) > at > org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:834) > at > org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala) > Caused by: org.apache.spark.rpc.RpcEnvStoppedException: RpcEnv already > stopped. > at > org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:158) > at > org.apache.spark.rpc.netty.Dispatcher.postLocalMessage(Dispatcher.scala:135) > at org.apache.spark.rpc.netty.NettyRpcEnv.ask(NettyRpcEnv.scala:229) > at > org.apache.spark.rpc.netty.NettyRpcEndpointRef.ask(NettyRpcEnv.scala:523) > at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:91) > ... 17 more > 18/03/07 23:34:22 INFO yarn.ApplicationMaster: Final app status: FAILED, > exitCode: 13, (reason: Uncaught exception: org.apache.spark.SparkException: > Exception thrown in awaitResult: ) > {code} > Example application: > {code:java} > object ExampleApp { > def main(args: Array[String]): Unit = { > val conf = new SparkConf().setAppName("ExampleApp") > val sc = new SparkContext(conf) > try { > // Do nothing > } finally { > sc.stop() > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-23660) Yarn throws exception in cluster mode when the application is small
[ https://issues.apache.org/jira/browse/SPARK-23660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16396388#comment-16396388 ] Gabor Somogyi commented on SPARK-23660: --- I'm working on it. > Yarn throws exception in cluster mode when the application is small > --- > > Key: SPARK-23660 > URL: https://issues.apache.org/jira/browse/SPARK-23660 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.3.0 >Reporter: Gabor Somogyi >Priority: Minor > > Yarn throws the following exception in cluster mode when the application is > really small: > {code:java} > 18/03/07 23:34:22 WARN netty.NettyRpcEnv: Ignored failure: > java.util.concurrent.RejectedExecutionException: Task > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask@7c974942 > rejected from > java.util.concurrent.ScheduledThreadPoolExecutor@1eea9d2d[Terminated, pool > size = 0, active threads = 0, queued tasks = 0, completed tasks = 0] > 18/03/07 23:34:22 ERROR yarn.ApplicationMaster: Uncaught exception: > org.apache.spark.SparkException: Exception thrown in awaitResult: > at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:205) > at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75) > at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92) > at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76) > at > org.apache.spark.deploy.yarn.YarnAllocator.(YarnAllocator.scala:102) > at > org.apache.spark.deploy.yarn.YarnRMClient.register(YarnRMClient.scala:77) > at > org.apache.spark.deploy.yarn.ApplicationMaster.registerAM(ApplicationMaster.scala:450) > at > org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:493) > at > org.apache.spark.deploy.yarn.ApplicationMaster.org$apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:345) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply$mcV$sp(ApplicationMaster.scala:260) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260) > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anon$5.run(ApplicationMaster.scala:810) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1920) > at > org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:809) > at > org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:259) > at > org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:834) > at > org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala) > Caused by: org.apache.spark.rpc.RpcEnvStoppedException: RpcEnv already > stopped. > at > org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:158) > at > org.apache.spark.rpc.netty.Dispatcher.postLocalMessage(Dispatcher.scala:135) > at org.apache.spark.rpc.netty.NettyRpcEnv.ask(NettyRpcEnv.scala:229) > at > org.apache.spark.rpc.netty.NettyRpcEndpointRef.ask(NettyRpcEnv.scala:523) > at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:91) > ... 17 more > 18/03/07 23:34:22 INFO yarn.ApplicationMaster: Final app status: FAILED, > exitCode: 13, (reason: Uncaught exception: org.apache.spark.SparkException: > Exception thrown in awaitResult: ) > {code} > Example application: > {code:java} > object ExampleApp { > def main(args: Array[String]): Unit = { > val conf = new SparkConf().setAppName("ExampleApp") > val sc = new SparkContext(conf) > try { > // Do nothing > } finally { > sc.stop() > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org