RE: how to specify cores and executor to run spark jobs simultaneously

2016-09-14 Thread 박경희
Hi Divya
I think, you did try to run spark jobs on yarn. And also I think, you would like to submit jobs to different queues on yarn for each. And maybe you need to prepare queues on yarn to run spark jobs by configuring scheduler.
Best regards,KyeongHee

- Original Message -
Sender : Divya Gehlot 
Date : 2016-09-14 15:08 (GMT+9)
Title : how to specify cores and executor to run spark jobs simultaneously

Hi,
I am on EMR cluster and My cluster configuration is as below:Number of nodes including master node - 3Memory:22.50 GBVCores Total : 16
Active Nodes : 2Spark version- 1.6.1
Parameter set in spark-default.conf








spark.executor.instances         2
spark.executor.cores             8
spark.driver.memory              10473M
spark.executor.memory            9658M
spark.default.parallelism        32




Would let me know if need any other info regarding the cluster .
The current configuration for spark-submit is --driver-memory 5G \--executor-memory 2G \--executor-cores 5 \--num-executors 10 \

Currently  with the above job configuration if I try to run another spark job it will be in accepted state till the first one finishes .How do I optimize or update the above spark-submit configurations to run some more spark jobs simultaneously 
Would really appreciate the help.
Thanks,Divya 

 


how to specify cores and executor to run spark jobs simultaneously

2016-09-14 Thread Divya Gehlot
Hi,

I am on EMR cluster and My cluster configuration is as below:
Number of nodes including master node - 3
Memory:22.50 GB
VCores Total : 16
Active Nodes : 2
Spark version- 1.6.1

Parameter set in spark-default.conf

spark.executor.instances 2
> spark.executor.cores 8
> spark.driver.memory  10473M
> spark.executor.memory9658M
> spark.default.parallelism32


Would let me know if need any other info regarding the cluster .

The current configuration for spark-submit is
--driver-memory 5G \
--executor-memory 2G \
--executor-cores 5 \
--num-executors 10 \


Currently  with the above job configuration if I try to run another spark
job it will be in accepted state till the first one finishes .
How do I optimize or update the above spark-submit configurations to run
some more spark jobs simultaneously

Would really appreciate the help.

Thanks,
Divya