I use importtsv to load data as HFile

hadoop jar hbase-0.92.1.jar importtsv
-Dimporttsv.bulk.output=/outputs/mytable.bulk
-Dimporttsv.columns=HBASE_ROW_KEY,ns: -Dimporttsv.separator=, mytable
/input

Then I use completebulkload to load those bulk data into my table

hadoop jar hbase-0.92.1.jar completebulkload /outputs/mytable.bulk mytable

However, the size of table is very huge (4.x GB). And it has only one
region. Oddly, why doesn't HBase split it into multiple regions? It did
exceed the size to split (256MB).

/hbase/mytable/71611409ea972a65b0876f953ad6377e/ns:

[image: enter image description here]

To split it, I try to use Split button on the Web UI of HBase. Sadly, it
shows

org.apache.hadoop.hbase.regionserver.CompactSplitThread: Region
mytable,,1334215360439.71611409ea972a65b0876f953ad6377e. not
splittable because midkey=null

I have more data to load. About 300GB, no matter how many data I have
loaded, it is still only one region. Also, it is still not splittable. Any
idea?

Reply via email to