[ 
https://issues.apache.org/jira/browse/PHOENIX-1437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14206728#comment-14206728
 ] 

Brian Johnson commented on PHOENIX-1437:
----------------------------------------

Bound per what? There are potentially several similar simultaneous queries 
running, some of which may be in the same process and some of which may be in 
other storm worker processes

> java.lang.OutOfMemoryError: unable to create new native thread
> --------------------------------------------------------------
>
>                 Key: PHOENIX-1437
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-1437
>             Project: Phoenix
>          Issue Type: Bug
>            Reporter: Taylor Finnell
>
> Getting a java.lang.OutOfMemoryError when using Phoenix on Storm. Here is the 
> full stack trace.
> {code}
> java.lang.OutOfMemoryError: unable to create new native thread
>       at java.lang.Thread.start0(Native Method) ~[na:1.7.0_45]
>       at java.lang.Thread.start(java/lang/Thread.java:713) ~[na:1.7.0_45]
>       at 
> java.util.concurrent.ThreadPoolExecutor.addWorker(java/util/concurrent/ThreadPoolExecutor.java:949)
>  ~[na:1.7.0_45]
>       at 
> java.util.concurrent.ThreadPoolExecutor.execute(java/util/concurrent/ThreadPoolExecutor.java:1360)
>  ~[na:1.7.0_45]
>       at 
> java.util.concurrent.AbstractExecutorService.submit(java/util/concurrent/AbstractExecutorService.java:132)
>  ~[na:1.7.0_45]
>       at 
> org.apache.phoenix.iterate.ParallelIterators.submitWork(org/apache/phoenix/iterate/ParallelIterators.java:356)
>  ~[stormjar.jar:na]
>       at 
> org.apache.phoenix.iterate.ParallelIterators.getIterators(org/apache/phoenix/iterate/ParallelIterators.java:265)
>  ~[stormjar.jar:na]
>       at 
> org.apache.phoenix.iterate.ConcatResultIterator.getIterators(org/apache/phoenix/iterate/ConcatResultIterator.java:44)
>  ~[stormjar.jar:na]
>       at 
> org.apache.phoenix.iterate.ConcatResultIterator.currentIterator(org/apache/phoenix/iterate/ConcatResultIterator.java:66)
>  ~[stormjar.jar:na]
>       at 
> org.apache.phoenix.iterate.ConcatResultIterator.next(org/apache/phoenix/iterate/ConcatResultIterator.java:86)
>  ~[stormjar.jar:na]
>       at 
> org.apache.phoenix.jdbc.PhoenixResultSet.next(org/apache/phoenix/jdbc/PhoenixResultSet.java:732)
>  ~[stormjar.jar:na]
>       at java.lang.reflect.Method.invoke(java/lang/reflect/Method.java:606) 
> ~[na:1.7.0_45]
>       at 
> RUBY.each(file:/mnt/hadoop/storm/supervisor/stormdist/korrelate_match_log_processor_staging_KOR-2325-online_sync_to_hbase_tf_part_three-1-1415715986/stormjar.jar!/lib/korrelate_match_log_processor/cleanroom_online_event_adapter.rb:51)
>  ~[na:na]
>       at 
> RUBY.finish_batch(file:/mnt/hadoop/storm/supervisor/stormdist/korrelate_match_log_processor_staging_KOR-2325-online_sync_to_hbase_tf_part_three-1-1415715986/stormjar.jar!/lib/korrelate_match_log_processor/bolt/abstract_event_reader_bolt.rb:68)
>  ~[na:na]
>       at 
> RUBY.finishBatch(/Users/tfinnell/.rvm/gems/jruby-1.7.11@O2O-jruby/gems/redstorm-0.6.6/lib/red_storm/proxy/batch_bolt.rb:51)
>  ~[na:na]
>       at 
> redstorm.proxy.BatchBolt.finishBatch(redstorm/proxy/BatchBolt.java:149) 
> ~[stormjar.jar:na]
>       at 
> redstorm.storm.jruby.JRubyTransactionalBolt.finishBatch(redstorm/storm/jruby/JRubyTransactionalBolt.java:56)
>  ~[stormjar.jar:na]
>       at 
> backtype.storm.coordination.BatchBoltExecutor.finishedId(backtype/storm/coordination/BatchBoltExecutor.java:76)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.coordination.CoordinatedBolt.checkFinishId(backtype/storm/coordination/CoordinatedBolt.java:259)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.coordination.CoordinatedBolt.execute(backtype/storm/coordination/CoordinatedBolt.java:322)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.daemon.executor$fn__4329$tuple_action_fn__4331.invoke(executor.clj:630)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.daemon.executor$fn__4329$tuple_action_fn__4331.invoke(backtype/storm/daemon/executor.clj:630)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.daemon.executor$mk_task_receiver$fn__4252.invoke(executor.clj:398)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.daemon.executor$mk_task_receiver$fn__4252.invoke(backtype/storm/daemon/executor.clj:398)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.disruptor$clojure_handler$reify__1747.onEvent(disruptor.clj:58)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.disruptor$clojure_handler$reify__1747.onEvent(backtype/storm/disruptor.clj:58)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.utils.DisruptorQueue.consumeBatchToCursor(backtype/storm/utils/DisruptorQueue.java:104)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.utils.DisruptorQueue.consumeBatchWhenAvailable(backtype/storm/utils/DisruptorQueue.java:78)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.disruptor$consume_batch_when_available.invoke(disruptor.clj:77)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.disruptor$consume_batch_when_available.invoke(backtype/storm/disruptor.clj:77)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.daemon.executor$fn__4329$fn__4341$fn__4388.invoke(executor.clj:745)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.daemon.executor$fn__4329$fn__4341$fn__4388.invoke(backtype/storm/daemon/executor.clj:745)
>  ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at backtype.storm.util$async_loop$fn__442.invoke(util.clj:436) 
> ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at 
> backtype.storm.util$async_loop$fn__442.invoke(backtype/storm/util.clj:436) 
> ~[storm-core-0.9.1.2.1.2.0-402.jar:0.9.1.2.1.2.0-402]
>       at clojure.lang.AFn.run(clojure/lang/AFn.java:24) 
> ~[clojure-1.4.0.jar:na]
>       at java.lang.Thread.run(java/lang/Thread.java:744) ~[na:1.7.0_45]
> {code}
> Here is some of the system configuration.
> {code}
> ulimit -a
> core file size          (blocks, -c) 0
> data seg size           (kbytes, -d) unlimited
> scheduling priority             (-e) 0
> file size               (blocks, -f) unlimited
> pending signals                 (-i) 240435
> max locked memory       (kbytes, -l) 64
> max memory size         (kbytes, -m) unlimited
> open files                      (-n) 1024
> pipe size            (512 bytes, -p) 8
> POSIX message queues     (bytes, -q) 819200
> real-time priority              (-r) 0
> stack size              (kbytes, -s) 10240
> cpu time               (seconds, -t) unlimited
> max user processes              (-u) 240435
> virtual memory          (kbytes, -v) unlimited
> file locks                      (-x) unlimited
> {code}
> None of the tables I am querying against have more than 4 regions.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to