How often does the index change? Can you use an HTTP cache and do this
once for each new index?

wunder

On 1/31/08 9:09 AM, "Andy Blower" <[EMAIL PROTECTED]> wrote:

> 
> Actually I do need all facets for a field, although I've just realised that
> the tests are limited to only 100. Ooops. So it should be worse in
> reality... erk.
> 
> Since that's what we do with our current search engine, Solr has to be able
> to compete with this. The fields are a mix of non-multi, non-tokenized and
> others which are. I've yet to experiment with this.
> 
> Thanks.
> 
> 
> shalinmangar wrote:
>> 
>> I can't give you a definitive answer based on the data you've provided.
>> However, do you really need to get *all* facets? Can't you limit them with
>> facet.limit field? Are you planning to run multiple *:* queries with all
>> facets turned on a 58GB index in a live system? I don't think that's a
>> good
>> idea.
>> 
>> As for the 125 seconds, I think it is probably because of paging issues.
>> Are
>> you faceting on multivalued or tokenized fields? In that case, Solr uses
>> field queries which consume a lot of memory if the number of unique terms
>> are large.
>> 
>> On Jan 31, 2008 9:13 PM, Andy Blower <[EMAIL PROTECTED]> wrote:
>> 
>>> 
>>> I'm evaluating SOLR/Lucene for our needs and currently looking at
>>> performance
>>> since 99% of the functionality we're looking for is provided. The index
>>> contains 18.4 Million records and is 58Gb in size. Most queries are
>>> acceptably quick, once the filters are cached. The filters select one or
>>> more of three subsets of the data and then intersect from around 15 other
>>> subsets of data depending on a user subscription.
>>> 
>>> We're returning facets on several fields, and sometimes a blank (q=*:*)
>>> query is run purely to get the facets for all of the data that the user
>>> can
>>> access. This information is turned into browse information and can be
>>> different for each user.
>>> 
>>> Running performance tests using jMeter sequentially with a single user,
>>> these blank queries are slower than the normal queries, but still in the
>>> 1-4sec range. Unfortunately if I increase the number of test threads so
>>> that
>>> more than one of the blank queries is submitted while one is already
>>> being
>>> processed, everything grinds to a halt and the responses to these blank
>>> queries can take up to 125secs to be returned!
>>> 
>>> This surprises me because the filter query submitted has usually already
>>> been submitted along with a normal query, and so should be cached in the
>>> filter cache. Surely all solr needs to do is return a handful of fields
>>> for
>>> the first 100 records in the list from the cache - or so I thought.
>>> 
>>> Can anyone tell me what might be causing this dramatic slowdown? Any
>>> suggestions for solutions would be gratefully received.
>>> 
>>> 
>>> Thans
>>> Andy.
>>> --
>>> View this message in context:
>>> http://www.nabble.com/Slow-response-times-using-*%3A*-tp15206563p15206563.ht
>>> ml
>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>> 
>>> 
>> 
>> 
>> -- 
>> Regards,
>> Shalin Shekhar Mangar.
>> 
>> 

Reply via email to