Hi, it need more information to clarify. can you paste your solrconfig.xml and OOO exception log?
On Mon, Oct 5, 2009 at 5:19 PM, Thomas Koch <tho...@koch.ro> wrote: > Hi, > > I'm running a read only index with SOLR 1.3 on a server with 8GB RAM and > the > Heap set to 6GB. The index contains 17 million documents and occupies 63GB > of > disc space with compression turned on. Replication frequency from the SOLR > master is 5 minutes. The index should be able to support around 10 > concurrent > searches. > > Now we start hitting RAM related errors like: > > - java.lang.OutOfMemoryError: Java heap space or > - java.lang.OutOfMemoryError: GC overhead limit exceeded > > which over time make the SOLR instance unresponsive. > > Before asking for advices on how to optimize my setup, I'd kindly ask for > your > experiences with setups of this size. Is it possible to run such a large > index > on only one server? Can I support even larger indexes when I tweak my > configuration? Where's the limit when I need to split the index on multiple > shards? When do I need to start considering a setup like/with Katta? > > Thanks for your insights, > > Thomas Koch, http://www.koch.ro >