The memory leak could be related to this
<https://issues.apache.org/jira/browse/SPARK-5967> defect that was resolved
in Spark 1.2.2 and 1.3.0.

It also was a HashMap causing the issue.

-Conor



On Wed, Apr 29, 2015 at 12:01 PM, Sean Owen <so...@cloudera.com> wrote:

> Please use user@, not dev@
>
> This message does not appear to be from your driver. It also doesn't say
> you ran out of memory. It says you didn't tell YARN to let it use the
> memory you want. Look at the memory overhead param and please search first
> for related discussions.
> On Apr 29, 2015 11:43 AM, "wyphao.2007" <wyphao.2...@163.com> wrote:
>
>> Hi, Dear developer, I am using Spark Streaming to read data from kafka,
>> the program already run about 120 hours, but today the program failed
>> because of driver's OOM as follow:
>>
>> Container [pid=49133,containerID=container_1429773909253_0050_02_000001]
>> is running beyond physical memory limits. Current usage: 2.5 GB of 2.5 GB
>> physical memory used; 3.2 GB of 50 GB virtual memory used. Killing
>> container.
>>
>> I set --driver-memory to 2g, In my mind, driver is responsibility for
>> job scheduler and job monitor(Please correct me If I'm wrong), Why it using
>> so much memory?
>>
>> So I using jmap to monitor other program(already run about 48 hours):
>> sudo /home/q/java7/jdk1.7.0_45/bin/jmap -histo:live 31256, the result as
>> follow:
>> the java.util.HashMap$Entry and java.lang.Long  object using about 600Mb
>> memory!
>>
>> and I also using jmap to monitor other program(already run about 1 hours
>> ),  the result as follow:
>> the java.util.HashMap$Entry and java.lang.Long object doesn't using so
>> many memory, But I found, as time goes by, the java.util.HashMap$Entry
>> and java.lang.Long object will occupied more and more memory,
>> It is driver's memory leak question? or other reason?
>>
>> Thanks
>> Best Regards
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>

Reply via email to