In addition to what Mike already said:

maxMergeDocs=99999 -- do you really mean maxMergeDocs and not maxBufferedDocs?

Larg(er) maxBufferedDocs will speed up indexing.

Otis 

--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch

----- Original Message ----
From: vivek sar <[EMAIL PROTECTED]>
To: java-user@lucene.apache.org
Sent: Friday, January 18, 2008 4:31:26 AM
Subject: Optimize for large index size

Hi,

  We are using Lucene 2.2. We have an index of size 70G (within 3-4
days) and growing. We run optimize pretty frequently (once every hour
- due to large number of index updates every min - can be up to 100K
new documents every min). I have seen every now and then the optimize
takes 3-4 hours to complete and up to 8 G memory (our limit). This
makes the whole system slow. Few questions,

1) Is there any alternative to optimize? That is, can we do without
optimize and still have our search fast?
2) What's the best way to use optimize, i.e. how can we make the
optimize much faster and use lesser memory?
3) Is there a way to partition the indexes using Lucene? Let's say we
partition daily, so we have to optimize only the daily indexes and not
the whole thing.

Our mergefactor=200 and maxMergeDocs=99999

Thanks,
-vivek

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]





---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to