Genady wrote:
Hi,
Just wondering if somebody could recommend a random read strategy for
searching a big group of the keys(100M) in hadoop/hbase cluster, using one
client is very slow, separating an input to smaller groups and running each
one with a different client is certainly improves performance, but maximum
speed I'm getting is ~3300 read/sec. I've tried to use map reduce and to run
search as map-reduce ask and to run HBase reads from map or reduce, but
HBase is start to fail. So hardware upgrade and creating HBase in memory
tables is only direction here?
Tell us more about your table schema, data sizes, and the types of
query. What performance do you need from hbase? Do your rows have many
columns and you are trying to get all columns when you query for
example? Are you on 0.19.0 Genady (sorry if you've answered this
question in the near past)?
St.Ack