Hi,
I did a search on the Lucene list archives, found a lot of posts about
the use of Lucene with NFS and how there are locking issues, but don't
see anybody coming to a real solution to this. Here's the most promising
thread I found:
Not exactly sure what you're asking with regards to Analyzers and
parsing XML...
But for parsing and indexing XML documents with Lucene, you can find a
lot of material out there by searching the list archives and using
google. However, the document I found most helpful was this piece
written by
Ok... just got confused because you mentioned XML. Unless you're
actually indexing the raw XML in some of your fields, the fact that
you're indexing XML documents as your source content is irrelevant to
your choice of Analyzer.
Choice of indexer really depends on your specific project
I just wanted to clarify... I don't believe the following statement is
accurate:
The contents field searches on the entire document,
including all indexes.
There is no default field named contents that automatically combines
the contents of all your fields. As Erik mentioned, you need to
You need to delete the document from the index and reindex it. This is
in the LuceneFAQ:
http://wiki.apache.org/jakarta-lucene/LuceneFAQ#head-917dd4fc904aa20a34e
bd23eb321125bdca1dea2
(or #24 under 3. Indexing)
Peter
-Original Message-
From: anushri kumar [mailto:[EMAIL PROTECTED]
Are you using a hit collector? I think if you use a hit collector rather
than the Hits object for getting query results, deleted items will still
be returned as results. My workaround for this was to run optimize after
I finish a batch of deletes, which works fine for my system because I
only run
The system I'm working on requires that I also need to use a custom hit
collector. I'm not using the RemoteSearchable. Instead, I'm just binding
an object called SearcherImpl that contains an IndexSearcher and a
getResults() method that uses the custom hit collector.
Peter
-Original
Hi,
I'm getting this error from trying to access a document in my custom hit
collector:
java.lang.IllegalArgumentException: attempt to access a deleted document
at
org.apache.lucene.index.SegmentReader.document(SegmentReader.java:235
)
at
I think my best option will just be to optimize the index after each
deletion. I guess this will be good for me anyways... For the cases
where optimizing is not a luxury one can afford, it would be nice if
there was another way around this.
Peter
-Original Message-
From: Peter Kim
is proportional to
the size of the index, is there any downside to splitting up a single
large index into up to 30 different indexes?
I would really appreciate your help.
Thanks!
Peter
-
Peter Kim
[EMAIL PROTECTED
Hi all,
I have a problem related to index size and deleting and optimizing. So
from reading various sources online, it seems as though the size of the
Lucene index should become no larger than half the size of the disk
since during optimization, the size of the index can ballon to double
the
11 matches
Mail list logo