On Tue, Nov 19, 2013 at 6:50 AM, Yadid Ayzenberg <[email protected]>wrote:

>  Hi all,
>
> According to the documentation, spark standalone currently only supports a
> FIFO scheduling system.
> I understand its possible to limit the number of cores a job uses by
> setting spark.cores.max.
> When running a job, will spark try using the max number of cores on each
> machine until it reaches the set limit, or will it do this round robin
> style - utilize a single core on each machine -  if its already used a core
> on all of the slaves and the limit has not been reached, spark will utilize
> an additional core on each machine and so on.
>
> I think the latter make more sense, but I want to be sure that is the case.
>
> Thanks,
> Yadid
>
>

Reply via email to