NFS/iSCSI SAN with Lucene

2006-07-19 Thread Peter Kim
Hi, I did a search on the Lucene list archives, found a lot of posts about the use of Lucene with NFS and how there are locking issues, but don't see anybody coming to a real solution to this. Here's the most promising thread I found:

RE: Analysis

2005-11-01 Thread Peter Kim
Not exactly sure what you're asking with regards to Analyzers and parsing XML... But for parsing and indexing XML documents with Lucene, you can find a lot of material out there by searching the list archives and using google. However, the document I found most helpful was this piece written by

RE: Analysis

2005-11-01 Thread Peter Kim
Ok... just got confused because you mentioned XML. Unless you're actually indexing the raw XML in some of your fields, the fact that you're indexing XML documents as your source content is irrelevant to your choice of Analyzer. Choice of indexer really depends on your specific project

RE: Help requested

2005-10-31 Thread Peter Kim
I just wanted to clarify... I don't believe the following statement is accurate: The contents field searches on the entire document, including all indexes. There is no default field named contents that automatically combines the contents of all your fields. As Erik mentioned, you need to

RE: Indexing

2005-10-31 Thread Peter Kim
You need to delete the document from the index and reindex it. This is in the LuceneFAQ: http://wiki.apache.org/jakarta-lucene/LuceneFAQ#head-917dd4fc904aa20a34e bd23eb321125bdca1dea2 (or #24 under 3. Indexing) Peter -Original Message- From: anushri kumar [mailto:[EMAIL PROTECTED]

RE: Delete doesn't delete?

2005-10-25 Thread Peter Kim
Are you using a hit collector? I think if you use a hit collector rather than the Hits object for getting query results, deleted items will still be returned as results. My workaround for this was to run optimize after I finish a batch of deletes, which works fine for my system because I only run

RE: RemoteSearchable woes

2005-10-13 Thread Peter Kim
The system I'm working on requires that I also need to use a custom hit collector. I'm not using the RemoteSearchable. Instead, I'm just binding an object called SearcherImpl that contains an IndexSearcher and a getResults() method that uses the custom hit collector. Peter -Original

IllegalArgumentException: attempt to access a deleted document

2005-10-06 Thread Peter Kim
Hi, I'm getting this error from trying to access a document in my custom hit collector: java.lang.IllegalArgumentException: attempt to access a deleted document at org.apache.lucene.index.SegmentReader.document(SegmentReader.java:235 ) at

RE: IllegalArgumentException: attempt to access a deleted document

2005-10-06 Thread Peter Kim
I think my best option will just be to optimize the index after each deletion. I guess this will be good for me anyways... For the cases where optimizing is not a luxury one can afford, it would be nice if there was another way around this. Peter -Original Message- From: Peter Kim

IndexSearcher architecture and memory usage

2005-09-21 Thread Peter Kim
is proportional to the size of the index, is there any downside to splitting up a single large index into up to 30 different indexes? I would really appreciate your help. Thanks! Peter - Peter Kim [EMAIL PROTECTED

Problem with deleting and optimizing index

2005-07-21 Thread Peter Kim
Hi all, I have a problem related to index size and deleting and optimizing. So from reading various sources online, it seems as though the size of the Lucene index should become no larger than half the size of the disk since during optimization, the size of the index can ballon to double the