Hi I have added more logging to my test application. I have two servers writing to a shared Lucene index on an NFS partition...
Here is the logging from one server... [10:49:18] [DEBUG] LuceneIndexAccessor closing cached writer [10:49:18] [DEBUG] ExpirationTimeDeletionPolicy onCommit() delete [segments_n] and the other server (at the same time): [10:49:18] [DEBUG] LuceneIndexAccessor opening new writer and caching it [10:49:18] [DEBUG] IndexAccessProvider getWriter() [10:49:18] [ERROR] DocumentCollection update(DocumentData) com.company.lucene.LuceneIcmException: I/O Error: Cannot add the document to the index. [/mnt/nfstest/repository/lucene/lucene-icm-test-1-0/segments_n (No such file or directory)] at com.company.lucene.RepositoryWriter.addDocument(RepositoryWriter.java:182) I think the exception is being thrown when the IndexWriter is created: new IndexWriter(directory, false, analyzer, false, deletionPolicy); I am confused... segments_n should not have been touched for 3 minutes so why would a new IndexWriter want to read it? Here is the whole of the stack trace: com.company.lucene.LuceneIcmException: I/O Error: Cannot add the document to the index. [/mnt/nfstest/repository/lucene/lucene-icm-test-1-0/segments_n (No such file or directory)] at com.company.lucene.RepositoryWriter.addDocument(RepositoryWriter.java:182) at com.company.lucene.IndexUpdate.addDocument(IndexUpdate.java:364) at com.company.lucene.IndexUpdate.addDocument(IndexUpdate.java:342) at com.company.lucene.IndexUpdate.update(IndexUpdate.java:67) at com.company.lucene.icm.DocumentCollection.update(DocumentCollection.java:390) at lucene.icm.test.Write.add(Write.java:105) at lucene.icm.test.Write.run(Write.java:79) at lucene.icm.test.Write.main(Write.java:43) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:324) at org.codehaus.mojo.exec.ExecJavaMojo$1.run(ExecJavaMojo.java:271) at java.lang.Thread.run(Thread.java:534) Caused by: java.io.FileNotFoundException: /mnt/nfstest/repository/lucene/lucene-icm-test-1-0/segments_n (No such file or directory) at java.io.RandomAccessFile.open(Native Method) at java.io.RandomAccessFile.<init>(RandomAccessFile.java:204) at org.apache.lucene.store.FSDirectory$FSIndexInput$Descriptor.<init>(FSDirectory.java:506) at org.apache.lucene.store.FSDirectory$FSIndexInput.<init>(FSDirectory.java:536) at org.apache.lucene.store.FSDirectory$FSIndexInput.<init>(FSDirectory.java:531) at org.apache.lucene.store.FSDirectory.openInput(FSDirectory.java:440) at org.apache.lucene.index.SegmentInfos.read(SegmentInfos.java:193) at org.apache.lucene.index.IndexFileDeleter.<init>(IndexFileDeleter.java:156) at org.apache.lucene.index.IndexWriter.init(IndexWriter.java:626) at org.apache.lucene.index.IndexWriter.<init>(IndexWriter.java:573) at com.subshell.lucene.indexaccess.impl.IndexAccessProvider.getWriter(IndexAccessProvider.java:68) at com.subshell.lucene.indexaccess.impl.LuceneIndexAccessor.getWriter(LuceneIndexAccessor.java:171) at com.company.lucene.RepositoryWriter.addDocument(RepositoryWriter.java:176) ... 13 more Thank you very much for your previous comments and emails. Any help solving this issue would be appreciated. Patrick On 30/06/07, Michael McCandless <[EMAIL PROTECTED]> wrote:
Patrick Kimber wrote: > I have been checking the application log. Just before the time when > the lock file errors occur I found this log entry: > [11:28:59] [ERROR] IndexAccessProvider > java.io.FileNotFoundException: > /mnt/nfstest/repository/lucene/lucene-icm-test-1-0/segments_h75 (No > such file or directory) > at java.io.RandomAccessFile.open(Native Method) I think this exception is the root cause. On hitting this IOException in reader.close(), that means this reader has not released its write lock. Is it possible to see the full stack trace? Having the wrong deletion policy or even a buggy deletion policy (if indeed file.lastModified() varies by too much across machines) can't cause this (I think). At worse, the wrong deletion policy should cause other already-open readers to hit "Stale NFS handle" IOExceptions during searching. So, you should use your ExpirationTimeDeletionPolicy when opening your readers if they will be doing deletes, but I don't think it explains this root-cause exception during close(). It's a rather spooky exception ... in close(), the reader initializes an IndexFileDeleter which lists the directory and opens any segments_N files that it finds. Do you have a writer on one machine closing, and then very soon thereafter this reader on a different machine does deletes and tries to close? My best guess is the exception is happening inside that initialization because the directory listing said that "segments_XXX" exists but then when it tries to open that file, it does not in fact exist. Since NFS client-side caching (especially directory listing cache) is not generally guaranteed to be "correct", it could explain this. But let's see the full stack trace to make sure this is it... Mike --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]
--------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]