Thanks...! Some questions below.
1) you are suggesting that maybe this OOME is a symptom/red herring , and the
true cause of it is that a thread can't span because of ulimit... If so
possibly this could be flagged early on in the build. And -- where are so many
threads coming from that I need to up my limit? Is this a new feature added
to spark recently, and if so will it effect deployments scenarios as well?
And
2) possibly SBT_OPTS is where the memory settings should be ? If so, then why
do we have the get_mem_opts wrapper function coded to send memory manually as
Xmx/Xms options?
execRunner "$java_cmd" \
${SBT_OPTS:-$default_sbt_opts} \
$(get_mem_opts $sbt_mem) \
${java_opts} \
${java_args[@]} \
-jar "$sbt_jar" \
"${sbt_commands[@]}" \
"${residual_args[@]}"
> On Aug 26, 2014, at 8:58 PM, Mubarak Seyed <[email protected]> wrote:
>
> What is your ulimit value?
>
>
>> On Tue, Aug 26, 2014 at 5:49 PM, jay vyas <[email protected]>
>> wrote:
>> Hi spark.
>>
>> I've been trying to build spark, but I've been getting lots of oome
>> exceptions.
>>
>> https://gist.github.com/jayunit100/d424b6b825ce8517d68c
>>
>> For the most part, they are of the form:
>>
>> java.lang.OutOfMemoryError: unable to create new native thread
>>
>> I've attempted to hard code the "get_mem_opts" function, which is in the
>> sbt-launch-lib.bash file, to
>> have various very high parameter sizes (i.e. -Xms5g") with high
>> MaxPermSize, etc... and to no avail.
>>
>> Any thoughts on this would be appreciated.
>>
>> I know of others having the same problem as well.
>>
>> Thanks!
>>
>> --
>> jay vyas
>