[ 
https://issues.apache.org/jira/browse/SOLR-1308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12735679#action_12735679
 ] 

Jason Rutherglen commented on SOLR-1308:
----------------------------------------

Perhaps in another issue we can implement a cache that is RAM
usage aware. Implement sizeof(bitset), and keep the cache below
a predefined limit? 

Do we need to have a cache per reader or can the cachekey
include the reader? If segments are created rapidly, we may not
want the overhead of creating a new cache and managing it's
size, especially if we move to a RAM usage model.  


> Cache docsets and docs at the SegmentReader level
> -------------------------------------------------
>
>                 Key: SOLR-1308
>                 URL: https://issues.apache.org/jira/browse/SOLR-1308
>             Project: Solr
>          Issue Type: Improvement
>    Affects Versions: 1.4
>            Reporter: Jason Rutherglen
>            Priority: Minor
>             Fix For: 1.5
>
>   Original Estimate: 504h
>  Remaining Estimate: 504h
>
> Solr caches docsets and documents at the top level Multi*Reader
> level. After a commit, the caches are flushed. Reloading the
> caches in near realtime (i.e. commits every 1s - 2min)
> unnecessarily consumes IO resources, especially for largish
> indexes.
> We can cache docsets and documents at the SegmentReader level.
> The cache settings in SolrConfig can be applied to the
> individual SR caches.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to