Thanks. 发件人: Naveen [mailto:hadoopst...@gmail.com] 发送时间: 2016年12月25日 0:33 收件人: Linyuxin <linyu...@huawei.com> 抄送: user <user@spark.apache.org> 主题: Re: 答复: submit spark task on yarn asynchronously via java?
Hi, Please use SparkLauncher API class and invoke the threads using async calls using Futures. Using SparkLauncher, you can mention class name, application resouce, arguments to be passed to the driver, deploy-mode etc. I would suggest to use scala's Future, is scala code is possible. https://spark.apache.org/docs/1.5.1/api/java/org/apache/spark/launcher/SparkLauncher.html https://docs.oracle.com/javase/7/docs/api/java/util/concurrent/Future.html On Fri, Dec 23, 2016 at 7:10 AM, Linyuxin <linyu...@huawei.com<mailto:linyu...@huawei.com>> wrote: Hi, Could Anybody help? 发件人: Linyuxin 发送时间: 2016年12月22日 14:18 收件人: user <user@spark.apache.org<mailto:user@spark.apache.org>> 主题: submit spark task on yarn asynchronously via java? Hi All, Version: Spark 1.5.1 Hadoop 2.7.2 Is there any way to submit and monitor spark task on yarn via java asynchronously?