Thank you very much J-D. I'll definitely try the patch.
Regarding the block cache, could you give me some details.
Is it controlled by /hfile.block.cache.size /parameter? I've checked
most of the parameters related to store&heap and this one I really don't
understand. For loading every block there is a disk access? If adding
data in a region(randomly), is the cache important or only for random
get/ scan and delete?
Thank you,
Iulia
On 04/22/2011 09:21 PM, Jean-Daniel Cryans wrote:
The datanodes don't consume much memory, we run ours with 1GB and give
the rest to the region servers.
BTW if you want to serve the whole dataset, depending on your SLA, you
might want to try HDFS-347 since concurrent HDFS access is rather
slow. The other choice would be to make sure you can hold everything
in the block cache so that means very little data per region server.
J-D
On Fri, Apr 22, 2011 at 2:17 AM, Iulia Zidaru<[email protected]> wrote:
Hi all,
Supposing we have to constantly hit all data stored, which is a good report
between the HDFS space used and the HBase heap size allocated per node? Do
you calculate it somehow?
Also, is there a report between the hadoop heap size and the hbase heap size
that we should take into consideration?
Thank you,
Iulia