Actually, you should probably not let your index grow beyond one-third the size 
of your disk.
 
a] You start of with your original index
b] During optimize, Lucene will initially write out files in non-compound file 
format.
c] Lucene will than combine the non-compound file format into the compound file 
format.
 
So before producing your final index, the disk usage can be three times the 
size of your final index. You can get around it by not trying to use the 
compound file format - but various people have mentioned other issues with 
having too many files open errors.
Lokesh

Peter Kim <[EMAIL PROTECTED]> wrote:
Hi all,

I have a problem related to index size and deleting and optimizing. So
from reading various sources online, it seems as though the size of the
Lucene index should become no larger than half the size of the disk
since during optimization, the size of the index can ballon to double
the original unoptimized size. Unfortunately, I've allowed my index to
grow past half the size of the disk (well, actually I only have 8% disk
capacity remaining). 

So, I can delete a chunk of documents by using IndexReader.delete() but
disk space won't actually be freed up until I run optimize(). But
running optimize() may cause the index to grow until the disk runs out
of space. 

What should I do? Should I just not optimize and believe that adding new
documents will automatically replace the space taken up by the deleted
documents? Or is my only option to nuke the index and remember next time
to not allow the index to grow to more than half the size of the disk?

Thanks,
Peter

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to