To add to the pile of information we're asking you to provide, what version
of Spark are you running?


2014-08-13 11:11 GMT-07:00 Shivaram Venkataraman <shiva...@eecs.berkeley.edu
>:

> If the JVM heap size is close to the memory limit the OS sometimes kills
> the process under memory pressure. I've usually found that lowering the
> executor memory size helps.
>
> Shivaram
>
>
> On Wed, Aug 13, 2014 at 11:01 AM, Matei Zaharia <matei.zaha...@gmail.com>
> wrote:
>
>> What is your Spark executor memory set to? (You can see it in Spark's web
>> UI at http://<driver>:4040 under the executors tab). One thing to be
>> aware of is that the JVM never really releases memory back to the OS, so it
>> will keep filling up to the maximum heap size you set. Maybe 4 executors
>> with that much heap are taking a lot of the memory.
>>
>> Persist as DISK_ONLY should indeed stream data from disk, so I don't
>> think that will be a problem.
>>
>> Matei
>>
>> On August 13, 2014 at 6:49:11 AM, rpandya (r...@iecommerce.com) wrote:
>>
>> After a lot of grovelling through logs, I found out that the Nagios
>> monitor
>> process detected that the machine was almost out of memory, and killed
>> the
>> SNAP executor process.
>>
>> So why is the machine running out of memory? Each node has 128GB of RAM,
>> 4
>> executors, about 40GB of data. It did run out of memory if I tried to
>> cache() the RDD, but I would hope that persist() is implemented so that
>> it
>> would stream to disk without trying to materialize too much data in RAM.
>>
>> Ravi
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Lost-executors-tp11722p12032.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>

Reply via email to