just for my clarification: off heap cannot be java objects, correct? so we
are always talking about serialized off-heap storage?
On May 20, 2014 1:27 AM, "Tathagata Das" <tathagata.das1...@gmail.com>
wrote:

> That's one the main motivation in using Tachyon ;)
> http://tachyon-project.org/
>
> It gives off heap in-memory caching. And starting Spark 0.9, you can cache
> any RDD in Tachyon just by specifying the appropriate StorageLevel.
>
> TD
>
>
>
>
> On Mon, May 19, 2014 at 10:22 PM, Mohit Jaggi <mohitja...@gmail.com>wrote:
>
>> I guess it "needs" to be this way to benefit from caching of RDDs in
>> memory. It would be nice however if the RDD cache can be dissociated from
>> the JVM heap so that in cases where garbage collection is difficult to
>> tune, one could choose to discard the JVM and run the next operation in a
>> few one.
>>
>>
>> On Mon, May 19, 2014 at 10:06 PM, Matei Zaharia 
>> <matei.zaha...@gmail.com>wrote:
>>
>>> They’re tied to the SparkContext (application) that launched them.
>>>
>>> Matei
>>>
>>> On May 19, 2014, at 8:44 PM, Koert Kuipers <ko...@tresata.com> wrote:
>>>
>>> from looking at the source code i see executors run in their own jvm
>>> subprocesses.
>>>
>>> how long to they live for? as long as the worker/slave? or are they tied
>>> to the sparkcontext and life/die with it?
>>>
>>> thx
>>>
>>>
>>>
>>
>

Reply via email to