Yes, I'm using 40%. i can't access to those data either. I don't know how YSCB executes the reads and if they are random and could take advange of the cache.
Do you think that it's an acceptable performance? 2015-01-29 16:26 GMT+01:00 Ted Yu <yuzhih...@gmail.com>: > What's the value for hfile.block.cache.size ? > > By default it is 40%. You may want to increase its value if you're using > default. > > Andrew published some ycsb results : > http://people.apache.org/~apurtell/results-ycsb-0.98.8/ycsb > -0.98.0-vs-0.98.8.pdf > > However, I couldn't access the above now. > > Cheers > > On Thu, Jan 29, 2015 at 7:14 AM, Guillermo Ortiz <konstt2...@gmail.com> > wrote: > >> Is there any result with that benchmark to compare?? >> I'm executing the different workloads and for example for 100% Reads >> in a table with 10Millions of records I only get an performance of >> 2000operations/sec. I hoped much better performance but I could be >> wrong. I'd like to know if it's a normal performance or I could have >> something bad configured. >> >> >> I have splitted the tabled and all the records are balanced and used >> snappy. >> The cluster has a master and 4 regions servers with 256Gb,Cores 2 (32 >> w/ Hyperthreading), 0.98.6-cdh5.3.0, >> >> RegionServer is executed with these parameters: >> /usr/java/jdk1.7.0_67-cloudera/bin/java -Dproc_regionserver >> -XX:OnOutOfMemoryError=kill -9 %p -Xmx1000m >> -Djava.net.preferIPv4Stack=true -Xms640679936 -Xmx640679936 >> -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:-CMSConcurrentMTEnabled >> -XX:CMSInitiatingOccupancyFraction=70 -XX:+CMSParallelRemarkEnabled >> -XX:OnOutOfMemoryError=/usr/lib64/cmf/service/common/killparent.sh >> -Dhbase.log.dir=/var/log/hbase >> >> -Dhbase.log.file=hbase-cmf-hbase-REGIONSERVER-cnsalbsrvcl23.lvtc.gsnet.corp.log.out >> -Dhbase.home.dir=/opt/cloudera/parcels/CDH-5.3.0-1.cdh5.3.0.p0.30/lib/hbase >> -Dhbase.id.str= -Dhbase.root.logger=INFO,RFA >> >> -Djava.library.path=/opt/cloudera/parcels/CDH-5.3.0-1.cdh5.3.0.p0.30/lib/hadoop/lib/native >> -Dhbase.security.logger=INFO,RFAS >> org.apache.hadoop.hbase.regionserver.HRegionServer start >> >> >> The results for 100% reads are >> [OVERALL], RunTime(ms), 42734.0 >> [OVERALL], Throughput(ops/sec), 2340.0570973931763 >> [UPDATE], Operations, 1.0 >> [UPDATE], AverageLatency(us), 103170.0 >> [UPDATE], MinLatency(us), 103168.0 >> [UPDATE], MaxLatency(us), 103171.0 >> [UPDATE], 95thPercentileLatency(ms), 103.0 >> [UPDATE], 99thPercentileLatency(ms), 103.0 >> [READ], Operations, 100000.0 >> [READ], AverageLatency(us), 412.5534 >> [READ], AverageLatency(us,corrected), 581.6249026771276 >> [READ], MinLatency(us), 218.0 >> [READ], MaxLatency(us), 268383.0 >> [READ], MaxLatency(us,corrected), 268383.0 >> [READ], 95thPercentileLatency(ms), 0.0 >> [READ], 95thPercentileLatency(ms,corrected), 0.0 >> [READ], 99thPercentileLatency(ms), 0.0 >> [READ], 99thPercentileLatency(ms,corrected), 0.0 >> [READ], Return=0, 100000 >> [CLEANUP], Operations, 1.0 >> [CLEANUP], AverageLatency(us), 103598.0 >> [CLEANUP], MinLatency(us), 103596.0 >> [CLEANUP], MaxLatency(us), 103599.0 >> [CLEANUP], 95thPercentileLatency(ms), 103.0 >> [CLEANUP], 99thPercentileLatency(ms), 103.0 >> >> hbase(main):030:0> describe 'username' >> DESCRIPTION >> ENABLED >> 'username', {NAME => 'cf', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER >> => 'ROW', REPLICATION_SCOPE => '0', true >> VERSIONS => '1', COMPRESSION => 'SNAPPY', MIN_VERSIONS => '0', TTL >> => 'FOREVER', KEEP_DELETED_CELLS => ' >> false', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'} >> 1 row(s) in 0.0170 seconds >> >> 2015-01-29 5:27 GMT+01:00 Ted Yu <yuzhih...@gmail.com>: >> > Maybe ask on Cassandra mailing list for the benchmark tool they use ? >> > >> > Cheers >> > >> > On Wed, Jan 28, 2015 at 1:23 PM, Guillermo Ortiz <konstt2...@gmail.com> >> > wrote: >> > >> >> I was checking that web, do you know if there's another possibility >> >> since last updated for Cassandra was two years ago and I'd like to >> >> compare bothof them with kind of same tool/code. >> >> >> >> 2015-01-28 22:10 GMT+01:00 Ted Yu <yuzhih...@gmail.com>: >> >> > Guillermo: >> >> > If you use hbase 0.98.x, please consider Andrew's ycsb repo: >> >> > >> >> > https://github.com/apurtell/ycsb/tree/new_hbase_client >> >> > >> >> > Cheers >> >> > >> >> > On Wed, Jan 28, 2015 at 12:41 PM, Nishanth S <nishanth.2...@gmail.com >> > >> >> > wrote: >> >> > >> >> >> You can use ycsb for this purpose.See here >> >> >> >> >> >> https://github.com/brianfrankcooper/YCSB/wiki/Getting-Started >> >> >> -Nishanth >> >> >> >> >> >> On Wed, Jan 28, 2015 at 1:37 PM, Guillermo Ortiz < >> konstt2...@gmail.com> >> >> >> wrote: >> >> >> >> >> >> > Hi, >> >> >> > >> >> >> > I'd like to do some benchmarks fo HBase but I don't know what tool >> >> >> > could use. I started to make some code but I guess that there're >> some >> >> >> > easier. >> >> >> > >> >> >> > I've taken a look to JMeter, but I guess that I'd attack directly >> from >> >> >> > Java, JMeter looks great but I don't know if it fits well in this >> >> >> > scenario. What tool could I use to take some measures as time to >> >> >> > response some read and write request, etc. I'd like that to be >> able to >> >> >> > make the same benchmarks to Cassandra. >> >> >> > >> >> >> >> >> >>