I have a doubt. When you process a Hbase table with MapReduce you could use
the TableInputFormat, I understand that it goes directly to HDFS files
(storesFiles in HDFS) , so you could do some filter in the map phase and
it's not the same to go through to the region servers to do some massive
queriesIt's possible to do the same using TableInputFormat with Spark and
it's more efficient than use scan with filters and so on (again) when you
want to do a massive query about all the table. Am I right?

Reply via email to