I have a doubt. When you process a Hbase table with MapReduce you could use the TableInputFormat, I understand that it goes directly to HDFS files (storesFiles in HDFS) , so you could do some filter in the map phase and it's not the same to go through to the region servers to do some massive queriesIt's possible to do the same using TableInputFormat with Spark and it's more efficient than use scan with filters and so on (again) when you want to do a massive query about all the table. Am I right?
- Reading the whole table with MapReduce and Spar... Guillermo Ortiz
- Re: Reading the whole table with MapReduce... sck
- Re: Reading the whole table with MapRe... Guillermo Ortiz Fernández
- Re: Reading the whole table with M... Guillermo Ortiz Fernández
- Re: Reading the whole table wi... James Kebinger