Maybe I don't understand, but if you have done the filter and extended tableinputformat, you can run a MR job with:
JobConf conf = new JobConf(...); conf.setInputFormat(YourTableInputFormat.class); Cheers, Tim On Tue, Apr 7, 2009 at 11:50 AM, Rakhi Khatwani <[email protected]> wrote: > Thanks Ryan, i will try that > > On Tue, Apr 7, 2009 at 3:05 PM, Ryan Rawson <[email protected]> wrote: > >> there is a server-side mechanism to filter rows, it's found in the >> org.apache.hadoop.hbase.filter package. im not sure how this interops with >> the TableInputFormat exactly. >> >> setting a filter to reduce the # of rows returned is pretty much exactly >> what you want. >> >> On Tue, Apr 7, 2009 at 2:26 AM, Rakhi Khatwani <[email protected] >> >wrote: >> >> > Hi, >> > i have a map reduce program with which i read from a hbase table. >> > In my map program i check if the column value of a is xxx, if yes then >> > continue with processing else skip it. >> > however if my table is really big, most of my time in the map gets wasted >> > for processing unwanted rows. >> > is there any way through which we could send a subset of rows (based on >> the >> > value of a particular column family) to the map??? >> > >> > i have also gone through TableInputFormatBase but am not able to figure >> out >> > how do we set the input format if we are using TableMapReduceUtil class >> to >> > initialize table map jobs. or is there any other way i could use it. >> > >> > Thanks in Advance, >> > Raakhi. >> > >> >
