hello all, I am still trying to find the best way to index a really big amount of data. at the moment I am trying to index each of the 29 textfiles in a single thread using for each an own IndexWriter and an own directory where to place the index. there are always six threads working the same time.
the problem that occures now is that every second thread stops due to a FileNotFoundException or an ArrayIndexOutOfBoundsException (the latter only once) while the other half finishes fine. the file's name is different for each thread but has always the extension ".fnm". for example: java.io.FileNotFoundException: /lucenetest/medlineIndex/1976-1977/_2zfj.fnm (Datei oder Verzeichnis nicht gefunden) at java.io.RandomAccessFile.open(Native Method) at java.io.RandomAccessFile.<init>(RandomAccessFile.java(Compiled Code)) at java.io.RandomAccessFile.<init>(RandomAccessFile.java(Compiled Code)) at org.apache.lucene.store.FSInputStream.<init>(Unknown Source) at org.apache.lucene.store.FSInputStream.<init>(Unknown Source) at org.apache.lucene.store.FSDirectory.openFile(Unknown Source) at org.apache.lucene.index.FieldInfos.<init>(Unknown Source) at org.apache.lucene.index.SegmentReader.<init>(Unknown Source) at org.apache.lucene.index.IndexWriter.mergeSegments(Unknown Source) at org.apache.lucene.index.IndexWriter.maybeMergeSegments(Unknown Source) at org.apache.lucene.index.IndexWriter.maybeMergeSegments(Unknown Source) at org.apache.lucene.index.IndexWriter.addDocument(Unknown Source) at de.biomax.lucenetest.MedlineRecordIndexer.indexDocs(MedlineRecordIndexer.java(Compiled Code)) since half of the files are indexed without throwing that kind of exception I'm at a loss where to start debugging. any ideas? thanks a lot chantal -- To unsubscribe, e-mail: <mailto:[EMAIL PROTECTED]> For additional commands, e-mail: <mailto:[EMAIL PROTECTED]>