I use hbase-0.98-5-hadoop2 and modify the default heap size of region server in hbase-env.sh as below (keep all the other parameters in the file default):
export HBASE_REGIONSERVER_OPTS="-Xmn200m -XX:CMSInitiatingOccupancyFraction=70 -Xms1024m -Xmx8000m" The error occurs when I start hbase cluster: 10.1.255.246: Invalid maximum heap size: -Xmx8000m 10.1.255.246: The specified size exceeds the maximum representable size. 10.1.255.246: Could not create the Java virtual machine. The jvm I use is 64 bit : java version "1.6.0_39" Java(TM) SE Runtime Environment (build 1.6.0_39-b04) Java HotSpot(TM) 64-Bit Server VM (build 20.14-b01, mixed mode) Why 8G setting exceeds the maximum representable size. 2014-09-15 11:39 GMT+08:00 Nick Dimiduk <ndimi...@gmail.com>: > The scripts automate use of the tool PerformanceEvaluation that ships with > HBase, so in that way it runs against a cluster directly. It depends on > having independent configuration directories set up for each test > config. There's probably too much custom-to-my-environment stuff in there, > but I hope I included enough diffs that you can work it out in your > deployment. Let me know if you have any more questions. > > -n > > On Sunday, September 14, 2014, 牛兆捷 <nzjem...@gmail.com> wrote: > > > Hi, Nick > > > > Can your perf_blockcache performance testing script can be applied to > hbase > > cluster directly? > > If not, what kind of things should I take care? > > > > 2014-08-22 7:06 GMT+08:00 Nick Dimiduk <ndimi...@gmail.com > <javascript:;> > > >: > > > > > I'm familiar with Stack's work too, but thanks for pointing it out :) > > > > > > On Wed, Aug 20, 2014 at 8:19 PM, 牛兆捷 <nzjem...@gmail.com > <javascript:;>> > > wrote: > > > > > > > Hi Nick: > > > > > > > > Yes, I am interested in it. I will try first. > > > > > > > > Btw, this site (http://people.apache.org/~stack/bc/) also does the > > > similar > > > > performance evaluation. > > > > You can have a look if you are interested in. > > > > > > > > > > > > 2014-08-21 1:48 GMT+08:00 Nick Dimiduk <ndimi...@gmail.com > > <javascript:;>>: > > > > > > > > > Hi Zhaojie, > > > > > > > > > > I'm responsible for this particular bit of work. One thing to note > in > > > > these > > > > > experiments is that I did not control explicitly for OS caching. I > > ran > > > > > "warmup" workloads before collecting measurements, but because the > > > amount > > > > > of RAM on the machine is fixed, it's impact of OS cache is > different > > > with > > > > > different based on the amount of memory used by HBase. Another, as > > Todd > > > > > pointed out on an earlier thread, is that my trend lines are > probably > > > > > optimistic/misleading. > > > > > > > > > > Something I was driving for was to understand how well the > different > > > > > implementations before as they're managing more and more memory. > I'd > > > like > > > > > to get some insight into how we might be able to take advantage of > > > 100's > > > > or > > > > > even 1000's of GB of memory when the time comes. That's part of why > > > > there's > > > > > so many variables. > > > > > > > > > > I scripted out the running of the tests, all of my configurations > are > > > > > available in the associated github repo [0], and all of the data > > points > > > > are > > > > > available as a csv. If you're interested in experimenting yourself, > > > > please > > > > > let me know how I can help. > > > > > > > > > > Cheers, > > > > > Nick > > > > > > > > > > [0]: https://github.com/ndimiduk/perf_blockcache > > > > > > > > > > > > > > > On Wed, Aug 20, 2014 at 6:00 AM, 牛兆捷 <nzjem...@gmail.com > > <javascript:;>> wrote: > > > > > > > > > > > the complete blog link is: > > > > > > http://zh.hortonworks.com/blog/blockcache-showdown-hbase/ > > > > > > > > > > > > > > > > > > 2014-08-20 11:41 GMT+08:00 牛兆捷 <nzjem...@gmail.com > <javascript:;> > > >: > > > > > > > > > > > > > Hi all: > > > > > > > > > > > > > > I saw some interesting results from Hortonworks blog (block > cache > > > > > > > < > > > > > > > > > > > > > > > > > > > > > http://zh.hortonworks.com/wp-content/uploads/2014/03/perfeval_blockcache_v2.pdf > > > > > > > > > > > > > > ). > > > > > > > > > > > > > > In this result, the ratio of memory footprint to database size > is > > > > held > > > > > > > fixed while > > > > > > > the absolute values are increased. > > > > > > > > > > > > > > In my mind, the performance should becomes worse for larger > ratio > > > as > > > > > the > > > > > > > increase > > > > > > > of the absolute value. For example BucketCache#(tmpfs), the > > > > difference > > > > > > > between ratio (DB"1.5":"RAM"1.0) and ratio (DB"4.5":"RAM"1.0) > > > becomes > > > > > > > larger as the increase of memory. > > > > > > > Actually, the result of ratio ( DB"1.5":"RAM"1.0) increase > > > linearly, > > > > > and > > > > > > > the result of ratio (DB"1.5":"RAM"1.0) exponentially. > > > > > > > > > > > > > > However, for BucketCache#(heap) and LruBlockCache, the result > is > > > out > > > > of > > > > > > my > > > > > > > expectation. > > > > > > > The curves of ratio (DB"1.5":"RAM"1.0) and ratio > > (DB"4.5":"RAM"1.0) > > > > > both > > > > > > > increase exponentially, but the relative differences as the > > > increase > > > > of > > > > > > > memory are not consistent. > > > > > > > Take LruBlockCache as an example, the difference of ratio > > > > > > > (DB"1.5":"RAM"1.0) and ratio (DB"4.5":"RAM"1.0) becomes smaller > > > from > > > > 20 > > > > > > GB > > > > > > > to 50 GB, but becomes larger from 50 GB to 60 GB. > > > > > > > > > > > > > > How to analysis the cause of this result, any ideas? > > > > > > > > > > > > > > -- > > > > > > > *Regards,* > > > > > > > *Zhaojie* > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > *Regards,* > > > > > > *Zhaojie* > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > *Regards,* > > > > *Zhaojie* > > > > > > > > > > > > > > > -- > > *Regards,* > > *Zhaojie* > > > -- *Regards,* *Zhaojie*