I'm not sure that we could parallelize it. Currently, its a serial process (as you say) - the queue collects across readers by adjusting the values in the queue to sort correctly against the current reader. That approach doesn't appear easily parallelized.

patrick o'leary wrote:
Think I may have found it, it was multiple runs of the filter, one for each segment reader, I was generating a new map to hold distances each time. So only the distances from the
last segment reader were stored.

Currently it looks like those segmented searches are done serially, well in solr they are-
I presume the end goal is to make them multi-threaded ?
I'll need to make my map synchronized


On Tue, Apr 28, 2009 at 4:42 PM, Uwe Schindler <[email protected] <mailto:[email protected]>> wrote:

    What is the problem exactly? Maybe you use the new Collector API,
    where the search is done for each segment, so caching does not
    work correctly?

    -----
    Uwe Schindler
    H.-H.-Meier-Allee 63, D-28213 Bremen
    http://www.thetaphi.de
    eMail: [email protected] <mailto:[email protected]>

    ------------------------------------------------------------------------

    *From:* patrick o'leary [mailto:[email protected]
    <mailto:[email protected]>]
    *Sent:* Tuesday, April 28, 2009 10:31 PM
    *To:* [email protected] <mailto:[email protected]>
    *Subject:* ReadOnlyMultiSegmentReader bitset id vs doc id

    hey

    I've got a filter that's storing document id's with a geo distance
    for spatial lucene using a bitset position for doc id,
    However with a MultiSegmentReader that's no longer going to working.

    What's the most appropriate way to go from bitset position to doc
    id now?

    Thanks
    Patrick




--
- Mark

http://www.lucidimagination.com




---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to