Hi all,

According to the documentation, spark standalone currently only supports a FIFO scheduling system. I understand its possible to limit the number of cores a job uses by setting |spark.cores.max. |When running a job, will spark try using the max number of cores on each machine until it reaches the set limit, or will it do this round robin style - utilize a single core on each machine - if its already used a core on all of the slaves and the limit has not been reached, spark will utilize an additional core on each machine and so on.

I think the latter make more sense, but I want to be sure that is the case.

Thanks,
Yadid

Reply via email to