Guillermo
How large is your table? How many columns? Sincerely, Sean > On May 28, 2019 at 10:11 AM Guillermo Ortiz <konstt2...@gmail.com > mailto:konstt2...@gmail.com > wrote: > > > I have a doubt. When you process a Hbase table with MapReduce you could > use > the TableInputFormat, I understand that it goes directly to HDFS files > (storesFiles in HDFS) , so you could do some filter in the map phase and > it's not the same to go through to the region servers to do some massive > queriesIt's possible to do the same using TableInputFormat with Spark and > it's more efficient than use scan with filters and so on (again) when you > want to do a massive query about all the table. Am I right? >