Hi,

we are using Apache Solr Cloud within a production environment. If the maximum heap-space is reached the Solr access time slows down, because of the working garbage collector for a small amount of time.

We use the following configuration:

- Apache Tomcat as webserver to run the Solr web application
- 13 indices with about 1500000 entries (300 MB)
- 5 server with one replication per index (5 GB max heap-space)
- All indices have the following caches
- maximum document-cache-size is 4096 entries, all other indices have between 64 and 1536 entries - maximum query-cache-size is 1024 entries, all other indices have between 64 and 768 - maximum filter-cache-size is 1536 entries, all other i ndices have between 64 and 1024
- the directory-factory-implementation is NRTCachingDirectoryFactory
- the index is updated once per hour (no auto commit)
- ca. 5000 requests per hour per server
- large filter-queries (up to 15000 bytes and 1500 boolean operations)
- many facet-queries (30%)

Behaviour:

Started with 512 MB heap space. Over several days the heap-space grow up, until the 5 GB was reached. At this moment the described problem occurs. From this time on the heap-space-useage is between 50 and 90 percent. No OutOfMemoryException occurs.

Questions:


1. Why does Solr use 5 GB ram, with this small amount of data?
2. Which impact does the large filter-queries have in relation to ram usage?

Thanks!

Johannes Siegert

Reply via email to