There's usually only a couple sort fields and a bunch of terms in the
various indices. The terms are user entered on various media so the
number of terms is very large.
Thanks for the help.
Todd
On 10/29/08, Todd Benge <[EMAIL PROTECTED]> wrote:
> Hi,
>
> I'm the lead engineer for search on a large website using lucene for search.
>
> We're indexing about 300M documents in ~ 100 indices. The indices add
> up to ~ 60G.
>
> The indices are sorted into 4 different Multisearcher with the largest
> handling ~50G.
>
> The code is basically like the following:
>
> private static MultiSearcher searcher;
>
> public void init(File files) {
>
> IndexSearcer [] searchers = new IndexSearcher[files.length] ();
> int i = 0;
> for ( File file: files ) {
> searchers[i++] = new IndexSearcher(FSDirectory.getDirectory(file);
> }
>
> searcher = new MultiSearcher(searchers);
> }
>
> public Searcher getSearcher() {
> return searcher;
> }
>
> We're seeing a high cache rate with Term & TermInfo in Lucene 2.4.
> Performance is good but servers are consistently hanging with
> OutOfMemory errors.
>
> We're allocating 4G in the heap to each server.
>
> Is there any way to control the amount of memory Lucene consume for
> caching? Any other suggestions on fixing the memory errors?
>
> Thanks,
>
> Todd
>
--
Sent from Gmail for mobile | mobile.google.com
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]