Hi,

   I am working on centos5 and I run elasticsearch with version 1.0.0 with 
*-Xms808m 
-Xmx808m -Xss256k*parameters. There are 17 index and total 30200583 docs. 
Each index's docs count between 1000000 and 2000000. I create request query 
like ( each index have date field );


{
  "query": {
    "bool": {
      "must": [
        {
          "range": {
            "date": {
              "to": "2014-06-01 14:14:00",
              "from": "2014-04-01 00:00:00"
            }
          }
        }
      ],
      "should": [],
      "must_not": [],
      "minimum_number_should_match": 1
    }
  },
  "from": 0,
  "size": "50"
}


It give response;

{
   took: 5903
   timed_out: false
   _shards: {
      total: 17
      successful: 17
      failed: 0
   },
   hits: {
   total: 30200583
...
...
...}



However when I send query on elasticsearch-head tool for last 50 rows like;

{
  ...
  ...
  ...
  "from": 30200533,
  "size": "50"
}

It does not give a response and throw exception like;

ava.lang.OutOfMemoryError: Java heap space
        at org.apache.lucene.store.DataOutput.copyBytes(DataOutput.java:247)
        at org.apache.lucene.store.Directory.copy(Directory.java:186)
        at 
org.elasticsearch.index.store.Store$StoreDirectory.copy(Store.java:348)
        at 
org.apache.lucene.store.TrackingDirectoryWrapper.copy(TrackingDirectoryWrapper.java:50)
        at 
org.apache.lucene.index.IndexWriter.createCompoundFile(IndexWriter.java:4596)
        at 
org.apache.lucene.index.DocumentsWriterPerThread.sealFlushedSegment(DocumentsWriterPerThread.java:535)
        at 
org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:502)
        at 
org.apache.lucene.index.DocumentsWriter.doFlush(DocumentsWriter.java:506)
        at 
org.apache.lucene.index.DocumentsWriter.flushAllThreads(DocumentsWriter.java:616)
        at org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:370)
        at 
org.apache.lucene.index.StandardDirectoryReader.doOpenFromWriter(StandardDirectoryReader.java:285)
        at 
org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:260)
        at 
org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:250)
        at 
org.apache.lucene.index.DirectoryReader.openIfChanged(DirectoryReader.java:170)
        at 
org.apache.lucene.search.XSearcherManager.refreshIfNeeded(XSearcherManager.java:123)
        at 
org.apache.lucene.search.XSearcherManager.refreshIfNeeded(XSearcherManager.java:59)
        at 
org.apache.lucene.search.XReferenceManager.doMaybeRefresh(XReferenceManager.java:180)
        at 
org.apache.lucene.search.XReferenceManager.maybeRefresh(XReferenceManager.java:229)
        at 
org.elasticsearch.index.engine.internal.InternalEngine.refresh(InternalEngine.java:730)
        at 
org.elasticsearch.index.shard.service.InternalIndexShard.refresh(InternalIndexShard.java:477)
        at 
org.elasticsearch.index.shard.service.InternalIndexShard$EngineRefresher$1.run(InternalIndexShard.java:924)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
        at java.lang.Thread.run(Thread.java:619)

What is the problem? Is it not enough java heap space or does my query 
cause this heap space error? 

I asked same question in stackoverflow. Soutions ,which recommended in 
stackoverflow, are not applicable for me. Anyone can give another solutions 
for this problem?

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/08a3e138-495b-4150-8520-06cfab04c31d%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to