Hi, I am using Lucene 2.4.1 to index a database with less than a million records. The resulting index is about 50MB in size. I keep getting an OutOfMemory Error if I re-use the same IndexWriter to index the complete database. This is though recommended in the performance hints. What I now do is, every 10000 Objects I close the index (and every 50 close actions optimize it) and create a new IndexWriter to continue. This process works fine, but to me seems hardly the recommended way to go. I've been using jhat/jmap as well as Netbeans profiler and am fairly sure that this is a problem related to Lucene.
Any Ideas - or post this to Jira ? Jira has quite a few OutOfMemory postings but they all seem closed in Version 2.4.1. Thanks, Stefan --------------------------------------------------------------------- To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org For additional commands, e-mail: java-user-h...@lucene.apache.org