The insert works well for when I have a row key which is constant for a long period of time, and I can split it up into blocks. But when the row key changes often, then insert performance over time starts to suffer. The suggestion made by Ryan does help, and I was eventually able to get the entire data set into HBase. ( ~120 Million records)
Currently working on some analysis, and had a question about the java api. Is there a way to get record count given a row key? something like: long getColumnCount (rowkey). So it doesn't bring down any data to client, but simply returns the size..? Thanks. stack-3 wrote: > > On Sat, Jun 6, 2009 at 6:13 PM, llpind <sonny_h...@hotmail.com> wrote: > >> >> >> And it's inserting 1M in about 1 minute+ . Not the best still. > > > What you looking for performance-wise? > > Is your cluster working for you now? > > Thanks, > St.Ack > > -- View this message in context: http://www.nabble.com/Frequent-changing-rowkey---HBase-insert-tp23906724p23926640.html Sent from the HBase User mailing list archive at Nabble.com.