I have not found a way to control the cores yet. This effectively limits
the cluster to a single application at a time. A subsequent application
shows in the 'WAITING' State on the dashboard.

On Wed, Oct 1, 2014 at 2:49 PM, Akshat Aranya <aara...@gmail.com> wrote:

>
>
> On Wed, Oct 1, 2014 at 11:33 AM, Akshat Aranya <aara...@gmail.com> wrote:
>
>>
>>
>> On Wed, Oct 1, 2014 at 11:00 AM, Boromir Widas <vcsub...@gmail.com>
>> wrote:
>>
>>> 1. worker memory caps executor.
>>> 2. With default config, every job gets one executor per worker. This
>>> executor runs with all cores available to the worker.
>>>
>>> By the job do you mean one SparkContext or one stage execution within a
>> program?  Does that also mean that two concurrent jobs will get one
>> executor each at the same time?
>>
>
> Experimenting with this some more, I figured out that an executor takes
> away "spark.executor.memory" amount of memory from the configured worker
> memory.  It also takes up all the cores, so even if there is still some
> memory left, there are no cores left for starting another executor.  Is my
> assessment correct? Is there no way to configure the number of cores that
> an executor can use?
>
>
>>
>>>
>>> On Wed, Oct 1, 2014 at 11:04 AM, Akshat Aranya <aara...@gmail.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> What's the relationship between Spark worker and executor memory
>>>> settings in standalone mode?  Do they work independently or does the worker
>>>> cap executor memory?
>>>>
>>>> Also, is the number of concurrent executors per worker capped by the
>>>> number of CPU cores configured for the worker?
>>>>
>>>
>>>
>>
>

Reply via email to