Do you think it would be a good idea to temper the use of CoProcessors? This kind of reminds me of when people first started using stored procedures...
Sent from a remote device. Please excuse any typos... Mike Segel On Nov 24, 2012, at 11:46 AM, tom <[email protected]> wrote: > Hi, but you do not need to us M/R. You could also use coprocessors. > > See this site: > https://blogs.apache.org/hbase/entry/coprocessor_introduction > -> in the section "Endpoints" > > An aggregation coprocessor ships with hbase that should match your > requirements. > You just need to load it and eventually you can access it from HTable: > > HTable.coprocessorExec(..) > <http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/HTable.html#coprocessorExec%28java.lang.Class,%20byte[],%20byte[],%20org.apache.hadoop.hbase.client.coprocessor.Batch.Call,%20org.apache.hadoop.hbase.client.coprocessor.Batch.Callback%29> > > Regards > tom > > Am 24.11.2012 18:32, schrieb Marcos Ortiz: >> Regards, Dalia. >> You have to use MapReduce for that. >> In the HBase in Practice´s book, there are lot of great examples for this. >> >> On 11/24/2012 12:15 PM, Dalia Sobhy wrote: >>> Dear all, >>> I wanted to ask a question.. >>> Do Hbase Aggregate Functions such as rowcount, getMax, get Average use >>> MapReduce to execute those functions? >>> Thanks :D >
