[ 
https://issues.apache.org/jira/browse/SOLR-5611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16048223#comment-16048223
 ] 

Justin Sarma edited comment on SOLR-5611 at 6/13/17 6:39 PM:
-------------------------------------------------------------

I've created an open source tool which you can use to approximate appropriate 
shards.rows values through a Monte Carlo simulation. You can specify what 
percentage of the time you want the results to be 100% accurate, for different 
search depths, and different shard counts. We used this tool at Shutterstock to 
improve Solr latency and reduce heap pressure:

https://github.com/jsarma/solr-shards-rows-optimizer

There's also a blog entry here which explains it in more detail, along with 
some perf tests of the results:

https://tech.shutterstock.com/2017/05/09/efficiently-handling-deep-pagination-in-a-distributed-search-engine/


was (Author: jsarma):
I've created a tool which you can use to approximate appropriate shards.rows 
values through a Monte Carlo simulation. You can specify what percentage of the 
time you want the results to be 100% accurate, for different search depths, and 
different shard counts. We used this tool at Shutterstock to improve Solr 
latency and reduce heap pressure:

https://github.com/jsarma/solr-shards-rows-optimizer

There's also a blog entry here which explains it in more detail, along with 
some perf tests of the results:

https://tech.shutterstock.com/2017/05/09/efficiently-handling-deep-pagination-in-a-distributed-search-engine/

> When documents are uniformly distributed over shards, enable returning 
> approximated results in distributed query
> ----------------------------------------------------------------------------------------------------------------
>
>                 Key: SOLR-5611
>                 URL: https://issues.apache.org/jira/browse/SOLR-5611
>             Project: Solr
>          Issue Type: Improvement
>          Components: SolrCloud
>            Reporter: Isaac Hebsh
>            Priority: Minor
>              Labels: distributed_search, shard, solrcloud
>             Fix For: 4.9, 6.0
>
>         Attachments: lec5-distributedIndexing.pdf
>
>
> Query with rows=1000, which sent to a collection of 100 shards (shard key 
> behaviour is default - based on hash of the unique key), will generate 100 
> requests of rows=1000, on each shard.
> This results to total number of rows*numShards unique keys to be retrieved. 
> This behaviour is getting worst as numShards grows.
> If the documents are uniformly distributed over the shards, the expected 
> number of document should be ~ rows/numShards. Obviously, there might be 
> extreme cases, when all of the top X documents are in a specific shard.
> I suggest adding an optional parameter, say approxResults=true, which decides 
> whether we should limit the rows in the shard requests to rows/numShardsor 
> not. Moreover, we can add a numeric parameter which increases the limit, to 
> be more accurate.
> For example, the query {{approxResults=true&approxResults.factor=1.5}} will 
> retrieve 1.5*rows/numShards from each shard. In the case of 100 shards and 
> rows=1000, each shard will return 15 documents.
> Furthermore, this can reduce the problem of deep paging, because the same 
> thing can be applied there. when requested start=100000, Solr creating shard 
> request with start=0 and rows=START+ROWS. In the approximated approach, start 
> parameter (in the shard requests) can be set to 100000/numShards. The idea of 
> the approxResults.factor creates some difficulties here, though.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Reply via email to