this should be enough to get started (you can pick the 1.* version if you
want the newer APIs and stuff, but for the elephant book, the older apis
will work fine as well) .
<dependencies>
    <dependency>
      <groupId>org.apache.hadoop</groupId>
      <artifactId>hadoop-core</artifactId>
      <version>0.20.2</version>
    </dependency>
</dependencies>


On Wed, Apr 24, 2013 at 3:13 PM, Kevin Burton <rkevinbur...@charter.net>wrote:

> I am reading “Hadoop in Action” and the author on page 51 puts forth this
> code:****
>
> ** **
>
> public class WordCount2 { ****
>
> public static void main(String[] args) { ****
>
>    JobClient client = new JobClient(); ****
>
>    JobConf conf = new JobConf(WordCount2.class); ****
>
>    FileInputFormat.addInputPath(conf, new Path(args[0]));****
>
>    FileOutputFormat.setOutputPath(conf, new Path(args[1])); ****
>
>    conf.setOutputKeyClass(Text.class);****
>
>    conf.setOutputValueClass(LongWritable.class);****
>
>    conf.setMapperClass(TokenCountMapper.class);****
>
>    conf.setCombinerClass(LongSumReducer.class);****
>
>    conf.setReducerClass(LongSumReducer.class);r****
>
>    client.setConf(conf);****
>
>    try {****
>
>        JobClient.runJob(conf);****
>
>    } catch (Exception e) {****
>
>        e.printStackTrace();****
>
>    } ****
>
>        } ****
>
> }****
>
> ** **
>
> Which is an example for a simple MapReduce job. But being a beginner I am
> not sure how to set up a project for this code. If I am using Maven what
> are the Maven dependencies that I need? There are several map reduce
> dependencies and I am not sure which to pick. Are there other dependencies
> need (such as JobConf)? What are the imports needed? During the
> construction of the configuration what heuristics are used to find the
> configuration for the Hadoop cluster?****
>
> ** **
>
> Thank you.****
>



-- 
Jay Vyas
http://jayunit100.blogspot.com

Reply via email to