Hi, I am working on centos5 and I run elasticsearch with version 1.0.0 with *-Xms808m -Xmx808m -Xss256k*parameters. There are 17 index and total 30200583 docs. Each index's docs count between 1000000 and 2000000. I create request query like ( each index have date field );
{ "query": { "bool": { "must": [ { "range": { "date": { "to": "2014-06-01 14:14:00", "from": "2014-04-01 00:00:00" } } } ], "should": [], "must_not": [], "minimum_number_should_match": 1 } }, "from": 0, "size": "50" } It give response; { took: 5903 timed_out: false _shards: { total: 17 successful: 17 failed: 0 }, hits: { total: 30200583 ... ... ...} However when I send query on elasticsearch-head tool for last 50 rows like; { ... ... ... "from": 30200533, "size": "50" } It does not give a response and throw exception like; ava.lang.OutOfMemoryError: Java heap space at org.apache.lucene.store.DataOutput.copyBytes(DataOutput.java:247) at org.apache.lucene.store.Directory.copy(Directory.java:186) at org.elasticsearch.index.store.Store$StoreDirectory.copy(Store.java:348) at org.apache.lucene.store.TrackingDirectoryWrapper.copy(TrackingDirectoryWrapper.java:50) at org.apache.lucene.index.IndexWriter.createCompoundFile(IndexWriter.java:4596) at org.apache.lucene.index.DocumentsWriterPerThread.sealFlushedSegment(DocumentsWriterPerThread.java:535) at org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:502) at org.apache.lucene.index.DocumentsWriter.doFlush(DocumentsWriter.java:506) at org.apache.lucene.index.DocumentsWriter.flushAllThreads(DocumentsWriter.java:616) at org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:370) at org.apache.lucene.index.StandardDirectoryReader.doOpenFromWriter(StandardDirectoryReader.java:285) at org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:260) at org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:250) at org.apache.lucene.index.DirectoryReader.openIfChanged(DirectoryReader.java:170) at org.apache.lucene.search.XSearcherManager.refreshIfNeeded(XSearcherManager.java:123) at org.apache.lucene.search.XSearcherManager.refreshIfNeeded(XSearcherManager.java:59) at org.apache.lucene.search.XReferenceManager.doMaybeRefresh(XReferenceManager.java:180) at org.apache.lucene.search.XReferenceManager.maybeRefresh(XReferenceManager.java:229) at org.elasticsearch.index.engine.internal.InternalEngine.refresh(InternalEngine.java:730) at org.elasticsearch.index.shard.service.InternalIndexShard.refresh(InternalIndexShard.java:477) at org.elasticsearch.index.shard.service.InternalIndexShard$EngineRefresher$1.run(InternalIndexShard.java:924) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:619) What is the problem? Is it not enough java heap space or does my query cause this heap space error? I asked same question in stackoverflow. Soutions ,which recommended in stackoverflow, are not applicable for me. Anyone can give another solutions for this problem? -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/08a3e138-495b-4150-8520-06cfab04c31d%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.