That's one the main motivation in using Tachyon ;)
http://tachyon-project.org/

It gives off heap in-memory caching. And starting Spark 0.9, you can cache
any RDD in Tachyon just by specifying the appropriate StorageLevel.

TD




On Mon, May 19, 2014 at 10:22 PM, Mohit Jaggi <mohitja...@gmail.com> wrote:

> I guess it "needs" to be this way to benefit from caching of RDDs in
> memory. It would be nice however if the RDD cache can be dissociated from
> the JVM heap so that in cases where garbage collection is difficult to
> tune, one could choose to discard the JVM and run the next operation in a
> few one.
>
>
> On Mon, May 19, 2014 at 10:06 PM, Matei Zaharia 
> <matei.zaha...@gmail.com>wrote:
>
>> They’re tied to the SparkContext (application) that launched them.
>>
>> Matei
>>
>> On May 19, 2014, at 8:44 PM, Koert Kuipers <ko...@tresata.com> wrote:
>>
>> from looking at the source code i see executors run in their own jvm
>> subprocesses.
>>
>> how long to they live for? as long as the worker/slave? or are they tied
>> to the sparkcontext and life/die with it?
>>
>> thx
>>
>>
>>
>

Reply via email to