On Jul 16, 2007, at 1:06 AM, Chris Hostetter wrote:
: Do we have a best practice for going from, say a SpanQuery doc/
: position information and retrieving the actual range of positions of
: content from the Document? Is it just to reanalyze the Document
: using the appropriate Analyzer and start recording once you hit the
: positions you are interested in? Seems like Term Vectors _could_
: help, but even my new Mapper approach patch (LUCENE-868) doesn't
: really help, because they are stored in a term-centric manner. I
: guess what I am after is a position centric approach. That is, give
this is kind of what i was suggesting in the last message i sent
to the java-user thread about paylods and SpanQueries (which i'm
guessing is what prompted this thread as well)...
http://www.nabble.com/Payloads-and-PhraseQuery-
tf3988826.html#a11551628
This is one use case, the other is related to the new patch I
submitted for LUCENE-960. In this case, I have a SpanQueryFilter
that identifies a bunch of docs and positions ahead of time. Then
the user enters new Span Query and I want to relate the matches from
the user query with the positions of matches in the filter and then
show that window.
my point was that currently, to retrieve a payload you need a
TermPositions instance, which is designed for iterating in the
order of...
seek(term)
skipTo(doc)
nextPosition()
getPayload()
...which is great for getting the payload of every instance
(ie:position) of a specific term in a given document (or in every
document) but without serious changes to the Spans API, the ideal
payload
API would let you say...
skipTo(doc)
advance(startPosition)
getPayload()
while (nextPosition() < endPosition)
getPosition()
but this seems like a nearly impossible API to implement given the
natore
of hte inverted index and the fact that terms aren't ever stored in
position order.
there's a lot i really don't know/understand about the lucene term
position internals ... but as i recall, the datastructure written
to disk
isn't actually a tree structure inverted index, it's a long
sequence of
tuples correct? so in theory you could scan along the tuples
untill you
find the doc you are interested in, ignoring all of the term info
along
the way, then whatever term you happen be on at the moment, you
could scan
along all of the positions until you find one in the range you are
interested in -- assuming you do, then you record the current Term
(and
read your payload data if interested)
I think the main issue I see is in both the payloads and the matching
case above is that they require a document centric approach. And
then, for each Document,
you ideally want to be able to just index into an array so that you
can go directly to the position that is needed based on Span.getStart()
if i remember correctly, the first part of this is easy, and
relative fast
-- i think skipTo(doc) on a TermDoc or TermPositions will happily
scan for
the first <term,doc> pair with the correct docId, irregardless of
the term
... the only thing i'm not sure about is how efficient it is to
loop over
nextPosition() for every term you find to see if any of them are in
your
range ... the best case scenerio is that the first position
returned is
above the high end of your range, in which case you can stop
immediately
and seek to the next term -- butthe worst case is that you call
nextPosition() over an over a lot of times before you get a
position in
(or above) your rnage .... an advancePosition(pos) that wokred like
seek
or skipTo might be helpful here.
: I feel like I am missing something obvious. I would suspect the
: highlighter needs to do this, but it seems to take the reanalyze
: approach as well (I admit, though, that I have little experience
with
: the highlighter.)
as i understand it the default case is to reanalyze, but if you have
TermFreqVector info stored with positions (ie: a
TermPositionVector) then
it can use that to construct a TokenStream by iterating over all
terms and
writing them into a big array in position order (see the
TermSources class
in the highlighter)
Ah, I see that now. Thanks.
this makes sense when highlighting because it doesn't know what
kind of
fragmenter is going to be used so it needs the whole TokenStream,
but it
seems less then ideal when you are only interested in a small
number of
position ranges that you know in advance.
: I am wondering if it would be useful to have an alternative Term
: Vector storage mechanism that was position centric. Because we
: couldn't take advantage of the lexicographic compression, it would
: take up more disk space, but it would be a lot faster for these
kinds
i'm not sure if it's really neccessary to store the data in a position
centric manner, assuming we have a way to "seek" by position like i
described above -- but then again i don't really know that what i
described above is all that possible/practical/performant.
I suppose I could use my Mapper approach to organize things in a
position centric way now that I think about it more. Just means some
unpacking and repacking. Still, probably would perform well enough
since I can setup the correct structure on the fly. I will give this
a try. Maybe even add a Mapper to do this.
-Grant
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]