There is a tableInputFormat class in 
org.apache.hadoop.hbase.mapreduce.TableInputFormat

Also, if you want to use TableMapReduceUtil you probably want to have your 
mapper function extend TableMapper.

Check out the javadocs for more info: 
http://hadoop.apache.org/hbase/docs/current/api/index.html



-----Original Message-----
From: Something Something [mailto:luckyguy2...@yahoo.com] 
Sent: Thursday, October 15, 2009 1:37 AM
To: gene...@hadoop.apache.org; hbase-user@hadoop.apache.org
Subject: Re: Question about MapReduce

If the answer is...

TableMapReduceUtil.initTableMapperJob

I apologize for the spam.  If this isn't the right way, please let me know.  
Thanks.


--- On Wed, 10/14/09, Something Something <luckyguy2...@yahoo.com> wrote:

From: Something Something <luckyguy2...@yahoo.com>
Subject: Question about MapReduce
To: gene...@hadoop.apache.org, hbase-user@hadoop.apache.org
Date: Wednesday, October 14, 2009, 10:18 PM

I would like to start a Map-Reduce job that does not read data from an input 
file or from a database.  I would like to pass 3 arguments to the Mapper & 
Reducer to work on.  Basically, these arguments are keys on the 3 different 
tables on HBase.

In other words, I don't want to use FileInputFormat or DbInputFormat because 
everything I need is already on HBase. 

How can I do this?  Please let me know.  Thanks.




      


      

Reply via email to