[ 
https://issues.apache.org/jira/browse/HBASE-15241?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack resolved HBASE-15241.
---------------------------
    Resolution: Invalid

Let me start over.

> Blockcache hits hbase.ui.blockcache.by.file.max limit and is silent that it 
> will load no more blocks 
> -----------------------------------------------------------------------------------------------------
>
>                 Key: HBASE-15241
>                 URL: https://issues.apache.org/jira/browse/HBASE-15241
>             Project: HBase
>          Issue Type: Sub-task
>          Components: BucketCache
>            Reporter: stack
>
> We can only load 100k blocks from a file. If 256Gs of SSD and blocks are 4k 
> in size to align with SSD block read, and you want it all in cache, the 100k  
> limit gets in the way (The 100k may be absolute limit... checking. In UI I 
> see 100k only). There is a configuration which lets you up the number per 
> file, hbase.ui.blockcache.by.file.max. This helps.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to