Hi Poonam, Could you say a bit about how you're doing this bulk load? Are you loading from a CSV file, or making JDBC calls yourself (or something else)?
Are you aware of the CSV bulk loader (http://phoenix.apache.org/bulk_dataload.html)? That will probably give you the best performance for a large load of data. - Gabriel On Wed, Sep 17, 2014 at 3:44 PM, Poonam Ligade <[email protected]> wrote: > Hi, > > I am trying to load around 15GB of data in hbase > Files are split in around 1.3GB. I have 2 regionservers and 1 master running > on seperate machines with 4 GB physical RAM each. > Heapsize For Hmaster and regionserver is 3 GB each > > How can I configure HBase to handle these writes faster. > > Currently I have below settings > > hbase.regionserver.handler.count 60 > hbase.client.write.buffer 8388608 > hbase.hregion.memstore.flush.size 134217728 > hbase.regionserver.global.memstore.lowerLimit 0.95 > > hbase.regionserver.global.memstore.upperLimit 0.4 > > hbase.hregion.max.filesize 3294967296 > > > > Regards, > Poonam
