Re: HBase bulk loaded region can't be splitted

2012-05-12 Thread Yifeng Jiang
Hi, You need to create your table with pre-split regions. $hbase org.apache.hadoop.hbase.util.RegionSplitter -c 10 -f region_name your_table This command will pre-create 10 regions in your table using MD5 strings as region boundaries. You can also customize the splitting algorithm. Please see

Re: HBase bulk loaded region can't be splitted

2012-05-12 Thread Subir S
Wouldn't major_compact trigger a split...if it really needs to split However if you want to presplit regions for your table you can use the regionsplitter utility as below: $export HADOOP_CLASSPATH=`hbase classpath`; hbase org.apache.hadoop.hbase.util.RegionSplitter This will give you a usag

Re: HBase bulk loaded region can't be splitted

2012-05-10 Thread Bruce Bian
Yes, I understand that. But after I complete the bulk load, shouldn't it trigger the region server to split that region in order to meet the *hbase*.*hregion*.*max*.*filesize * criteria? When I try to split the regions manually using the WebUI, nothing happened, but instead a Region mytable,,13342

Re: HBase bulk loaded region can't be splitted

2012-05-10 Thread Bryan Beaudreault
I haven't done bulk loads using the importtsv tool, but I imagine it works similarly to the mapreduce bulk load tool we are provided. If so, the following stands. In order to do a bulk load you need to have a table ready to accept the data. The bulk load does not create regions, but only puts da

HBase bulk loaded region can't be splitted

2012-05-10 Thread Bruce Bian
I use importtsv to load data as HFile hadoop jar hbase-0.92.1.jar importtsv -Dimporttsv.bulk.output=/outputs/mytable.bulk -Dimporttsv.columns=HBASE_ROW_KEY,ns: -Dimporttsv.separator=, mytable /input Then I use completebulkload to load those bulk data into my table hadoop jar hbase-0.92.1.jar com

HBase bulk loaded region can't be splitted

2012-05-10 Thread Bruce Bian
I use importtsv to load data as HFile hadoop jar hbase-0.92.1.jar importtsv -Dimporttsv.bulk.output=/outputs/mytable.bulk -Dimporttsv.columns=HBASE_ROW_KEY,ns: -Dimporttsv.separator=, mytable /input Then I use completebulkload to load those bulk data into my table hadoop jar hbase-0.92.1.jar com