Re: reduce more than one way

2008-11-09 Thread Owen O'Malley


On Nov 7, 2008, at 12:35 PM, Elia Mazzawi wrote:

I have 2 hadooop map/reduce programs that have the same map, but a  
different reduce methods.


can i run them in a way so that the map only happens once?


If the input to the reduces is the same, you can put the two reduces  
together and use one of the multiple output libraries. That will let  
your reducer produce two different output directories.


-- Owen


Re: reduce more than one way

2008-11-09 Thread Amar Kamat

Elia Mazzawi wrote:

Hello,

I'm writing hadoop programs in Java,
I have 2 hadooop map/reduce programs that have the same map, but a 
different reduce methods. 
Look how MultipleOutputFormat is used. This provides the facility to 
write to multiple files.

Amar


can i run them in a way so that the map only happens once?

maybe store the intermediate result or something?




Re: reading input for a map function from 2 different files?

2008-11-09 Thread Amar Kamat

some speed wrote:

I was wondering if it was possible to read the input for a map function from
2 different files:
  
1st file --- user-input file from a particular location(path)

2nd file=--- A resultant file (has just one key,value pair) from a
previous MapReduce job. (I am implementing a chain MapReduce function)

Now, for every key,value pair in the user-input file, I would like to use
the same key,value pair from the 2nd file for some calculations.
  
I guess you might need to do some kind of join on the 2 files. Look at 
contrib/data_join for more details.

Amar

Is it possible for me to do so? Can someone guide me in the right direction
please?


Thanks!

  




hadoop with tomcat

2008-11-09 Thread ZhiHong Fu
Hello:

  I have implemented a Map/Reduce job, which will receive data from
several web services and deal with it based on hadoop. But I want to build a
web application to manange this web service and Monitor the Map/Reduce job
process, which will be deployed with Tomcat , I have read several mails
about the hadoop with tomcat  , But I can't still get it going. I am very
confused.
 Can anybody give me an example or where I can find  something about it
in much more details, Thanks.


Re: error with inout for the Reduce method

2008-11-09 Thread Per Jacobsson
Change the name of the reduce method to be all lower case -- public void
reduce(...

Right now the compiler is complaining that you haven't overridden the
correct abstract method in the base class.
/ Per

On Sat, Nov 8, 2008 at 10:44 PM, pols cut [EMAIL PROTECTED] wrote:

   I am trying to get a simple map reduce function going but getting the
 following error. could anyone please help/guide me

 filename.java is not abstract and does not override abstract method
 reduceorg.apache.io.text,java.util.iteratororg.apache.hadoop.io.text,org.apache.hadoop.mapred.OutputCollectororg.apache.hadoop.io.Text.org.apache.hadoop.mapred.Reporter
 in org.apache.hadoop.mapred.Reducer
 public static class Reduce extends MapReduceBase implements
 ReducerText,Text,Text,Text


 My Map method is like this:
 public static class map extends MapReduceBase implements
 MapperLongWritable, Text, Text, Text
{
 public void map(LongWritable key, Text value, OutputCollectorText, Text
 output, Reporter reporter) throws IOException
{
  output.collect(word_key,new Text(str1));  //output.collect(Text,Text)

}

}

 public static class Reduce extends MapReduceBase implements ReducerText,
 Text, Text, Text
{

public void Reduce(Text key, IteratorText values,
 OutputCollectorText, Text output, Reporter reporter) throws IOException
   {
  output.collect(new Text(Test), new Text(Me));

  }
   }

 configuration is like :

conf.setMapOutputKeyClass(Text.class);
conf.setMapOutputValueClass(Text.class);
conf.setOutputKeyClass(Text.class);
conf.setOutputValueClass(Text.class);

conf.setMapperClass(map.class);
conf.setCombinerClass(Reduce.class);
conf.setReducerClass(Reduce.class);

conf.setInputFormat(TextInputFormat.class);
conf.setOutputFormat(TextOutputFormat.class);


  Add more friends to your messenger and enjoy! Go to
 http://messenger.yahoo.com/invite/


Re: reading input for a map function from 2 different files?

2008-11-09 Thread Amareshwari Sriramadasu

some speed wrote:

I was wondering if it was possible to read the input for a map function from
2 different files:

1st file --- user-input file from a particular location(path)
2nd file=--- A resultant file (has just one key,value pair) from a
previous MapReduce job. (I am implementing a chain MapReduce function)

Now, for every key,value pair in the user-input file, I would like to use
the same key,value pair from the 2nd file for some calculations.

  
I think you can use DistributedCache for distributing your second file 
among maps.
Please see more documentation at 
http://hadoop.apache.org/core/docs/current/mapred_tutorial.html#DistributedCache


Thanks
Amareshwari

Is it possible for me to do so? Can someone guide me in the right direction
please?


Thanks!

  




reading input for a map function from 2 different files?

2008-11-09 Thread some speed
I was wondering if it was possible to read the input for a map function from
2 different files:

1st file --- user-input file from a particular location(path)
2nd file=--- A resultant file (has just one key,value pair) from a
previous MapReduce job. (I am implementing a chain MapReduce function)

Now, for every key,value pair in the user-input file, I would like to use
the same key,value pair from the 2nd file for some calculations.

Is it possible for me to do so? Can someone guide me in the right direction
please?


Thanks!


Re: Question on opening file info from namenode in DFSClient

2008-11-09 Thread stack

Taeho Kang wrote:

Hi, thanks for your reply Dhruba,

One of my co-workers is writing a BigTable-like application that could be
used for online, near-real-time, services. 
Can your co-worker be convinced to instead spend his time helping-along 
the ongoing bigtable-like efforts?

I think HBase developers would have run into similar issues as well.
  
In hbase, we open the file once and keep it open.  File is shared 
amongst all clients.


St.Ack