The insert works well for when I have a row key which is constant for a long
period of time, and I can split it up into blocks.  But when the row key
changes often, then insert performance over time starts to suffer.  The
suggestion made by Ryan does help, and I was eventually able to get the
entire data set into HBase. ( ~120 Million records)

Currently working on some analysis, and had a question about the java api. 
Is there a way to get record count given a row key?  something like: long
getColumnCount (rowkey).  So it doesn't bring down any data to client, but
simply returns the size..?

Thanks.



stack-3 wrote:
> 
> On Sat, Jun 6, 2009 at 6:13 PM, llpind <sonny_h...@hotmail.com> wrote:
> 
>>
>>
>> And it's inserting 1M in about 1 minute+ .   Not the best still.
> 
> 
> What you looking for performance-wise?
> 
> Is your cluster working for you now?
> 
> Thanks,
> St.Ack
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Frequent-changing-rowkey---HBase-insert-tp23906724p23926640.html
Sent from the HBase User mailing list archive at Nabble.com.

Reply via email to