Sampling has been on my TODO list for the JSON Facet API. How much it would help depends on where the bottlenecks are, but that in conjunction with a hashing approach to collection (assuming field cardinality is high) should definitely help.
-Yonik On Fri, Nov 4, 2016 at 3:02 PM, John Davis <johndavis925...@gmail.com> wrote: > Hi, > I am trying to improve the performance of queries with facets. I understand > that for queries with high facet cardinality and large number results the > current facet computation algorithms can be slow as they are trying to loop > across all docs and facet values. > > Does there exist an option to compute facets by just looking at the top-n > results instead of all of them or a sample of results based on some query > parameters? I couldn't find one and if it does not exist, has this come up > before? This would definitely not be a precise facet count but using > reasonable sampling algorithms we should be able to extrapolate well. > > Thank you in advance for any advice! > > John