Maybe I don't understand, but if you have done the filter and extended
tableinputformat, you can run a MR job with:

JobConf conf = new JobConf(...);
conf.setInputFormat(YourTableInputFormat.class);

Cheers,

Tim




On Tue, Apr 7, 2009 at 11:50 AM, Rakhi Khatwani
<[email protected]> wrote:
> Thanks Ryan, i will try that
>
> On Tue, Apr 7, 2009 at 3:05 PM, Ryan Rawson <[email protected]> wrote:
>
>> there is a server-side mechanism to filter rows, it's found in the
>> org.apache.hadoop.hbase.filter package.  im not sure how this interops with
>> the TableInputFormat exactly.
>>
>> setting a filter to reduce the # of rows returned is pretty much exactly
>> what you want.
>>
>> On Tue, Apr 7, 2009 at 2:26 AM, Rakhi Khatwani <[email protected]
>> >wrote:
>>
>> > Hi,
>> >     i have a map reduce program with which i read from a hbase table.
>> > In my map program i check if the column value of a is xxx, if yes then
>> > continue with processing else skip it.
>> > however if my table is really big, most of my time in the map gets wasted
>> > for processing unwanted rows.
>> > is there any way through which we could send a subset of rows (based on
>> the
>> > value of a particular column family) to the map???
>> >
>> > i have also gone through TableInputFormatBase but am not able to figure
>> out
>> > how do we set the input format if we are using TableMapReduceUtil class
>> to
>> > initialize table map jobs. or is there any other way i could use it.
>> >
>> > Thanks in Advance,
>> > Raakhi.
>> >
>>
>

Reply via email to