[EMAIL PROTECTED] wrote:
Any chance of your figuring where the time is being spent? Do you have
the facility to add dumb logging, recompile, restart cluster, and then
check? If you want me to send you a patch to get you going, just say so.
i have some hypothesis, but it's only my speculation :
i think that, when i call scanner.iterator().hasNext() it has to check
all rows in the table to find (or not) next rowId (because hasnext time
rise in line correlation with a table size). Maby problem is in sorting
rowId in the table?
Scanners march step through the memcache and files in the filesystem.
When you call hasNext, open iterators are moved to the next row. There
should not be full region/table scanning going on per hasNext call.
Are you using filters per chance?
i will try to change rowId format.
i try checkout latest version of hbase and try to check the same case
today.
Thanks for doing this investigation. Which version are you currently
on? 0.1.2?
St.Ack