This is not a storm issue.
https://plumbr.eu/outofmemoryerror/unable-to-create-new-native-thread

Thank you for your time!

+++++++++++++++++++++
Jeff Maass <maas...@gmail.com>
linkedin.com/in/jeffmaass
stackoverflow.com/users/373418/maassql
+++++++++++++++++++++


On Tue, May 19, 2015 at 1:51 PM, saiprasad mishra <saiprasadmis...@gmail.com
> wrote:

> It has nothing to do with the heap as the link suggested
>
> Do a ulimit -a and see whats the value for
>
> max user processes
>
>
> Try to increase this if you are on linux
>
>
> Also do not forget to increase the open files too
>
>
> May be set both of them to 65535 and that should take care of this issue
>
>
> Regards
>
> Sai
>
> On Tue, May 19, 2015 at 11:09 AM, Ganesh Chandrasekaran <
> gchandraseka...@wayfair.com> wrote:
>
>>  Hi all,
>>
>> I am seeing the following error in one of the supervisor nodes which is
>> similar to the following question on stackoverflow –
>> http://stackoverflow.com/questions/23008467/java-lang-outofmemoryerror-unable-to-create-new-native-thread-while-running-s
>>
>>
>>
>> java.lang.OutOfMemoryError: unable to create new native thread at
>> java.lang.Thread.start0(Native Method) at
>> java.lang.Thread.start(Thread.java:691) at
>> java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:943)
>> at
>> java.util.concurrent.ThreadPoolExecutor.ensurePrestart(ThreadPoolExecutor.java:1555)
>> at
>> java.util.concurrent.ScheduledThreadPoolExecutor.delayedExecute(ScheduledThreadPoolExecutor.java:333)
>> at
>> java.util.concurrent.ScheduledThreadPoolExecutor.scheduleAtFixedRate(ScheduledThreadPoolExecutor.java:570)
>> at
>> java.util.concurrent.Executors$DelegatedScheduledExecutorService.scheduleAtFixedRate(Executors.java:695)
>> at backtype.storm.task.ShellBolt.prepare(ShellBolt.java:128) at
>> backtype.storm.daemon.executor$fn__4722$fn__4734.invoke(executor.clj:692)
>> at backtype.storm.util$async_loop$fn__458.invoke(util.clj:461) at
>> clojure.lang.AFn.run(AFn.java:24) at java.lang.Thread.run(Thread.java:722)
>>
>>
>>
>> I checked the memory on my supervisor and it still has 40% of memory
>> free. I am not sure if it is the same problem as the one discussed in the
>> link above. Also, I am only seeing this on 1 supervisor and not on my other
>> supervisor nodes. It was the same config we were having on our old cluster
>> that was running 0.8.2. I also have multiple topologies running on the
>> cluster but I doubt any application is causing memory leak. If it was
>> shouldn’t all memory be used up and not list anything as free?
>>
>>
>>
>> Cluster summary –
>>
>> 1 Nimbus - 5 Supervisors (12 ports each)
>>
>>
>>
>> drpc.childopts
>>
>> -Xmx768m
>>
>>
>>
>>
>>
>>
>>
>> nimbus.childopts
>>
>> -Xmx1024m
>>
>>
>>
>> supervisor.childopts
>>
>> -Xmx256m
>>
>>
>>
>> worker.childopts
>>
>> -Xmx1024m
>>
>>
>>
>> Thanks,
>>
>> Ganesh
>>
>>
>>
>>
>>
>
>

Reply via email to