this should be enough to get started (you can pick the 1.* version if you
want the newer APIs and stuff, but for the elephant book, the older apis
will work fine as well) .
<dependencies>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-core</artifactId>
<version>0.20.2</version>
</dependency>
</dependencies>
On Wed, Apr 24, 2013 at 3:13 PM, Kevin Burton <[email protected]>wrote:
> I am reading “Hadoop in Action” and the author on page 51 puts forth this
> code:****
>
> ** **
>
> public class WordCount2 { ****
>
> public static void main(String[] args) { ****
>
> JobClient client = new JobClient(); ****
>
> JobConf conf = new JobConf(WordCount2.class); ****
>
> FileInputFormat.addInputPath(conf, new Path(args[0]));****
>
> FileOutputFormat.setOutputPath(conf, new Path(args[1])); ****
>
> conf.setOutputKeyClass(Text.class);****
>
> conf.setOutputValueClass(LongWritable.class);****
>
> conf.setMapperClass(TokenCountMapper.class);****
>
> conf.setCombinerClass(LongSumReducer.class);****
>
> conf.setReducerClass(LongSumReducer.class);r****
>
> client.setConf(conf);****
>
> try {****
>
> JobClient.runJob(conf);****
>
> } catch (Exception e) {****
>
> e.printStackTrace();****
>
> } ****
>
> } ****
>
> }****
>
> ** **
>
> Which is an example for a simple MapReduce job. But being a beginner I am
> not sure how to set up a project for this code. If I am using Maven what
> are the Maven dependencies that I need? There are several map reduce
> dependencies and I am not sure which to pick. Are there other dependencies
> need (such as JobConf)? What are the imports needed? During the
> construction of the configuration what heuristics are used to find the
> configuration for the Hadoop cluster?****
>
> ** **
>
> Thank you.****
>
--
Jay Vyas
http://jayunit100.blogspot.com