, you can load specific jars in your
mapreduce tasks. But I'm not sure why you'd need to do this, since you
should be able to include any new code in the job jar you submit to hadoop.
-Michael
On 4/18/07 11:23 AM, Pedro Guedes [EMAIL PROTECTED] wrote:
I keep talking to myself... hope
Hi hadoopers,
I'm working on an enterprise search engine that works on an hadoop
cluster but is controlled form the outside. I managed to implement a
simple crawler much like Nutch's...
Now i have a new system's requirement: the crawl process must be
configurable outside hadoop. This means that I
I keep talking to myself... hope it doesn't annoy u too much!
We thought of a solution to our problem in wich we build a new .job
file, in accordance with our crawl configuration, and then pass it to
hadoop for execution... Is there somewhere i can look for the
specification of the .job format?
:06,285 INFO org.apache.hadoop.mapred.TaskTracker:
task_0001_r_01_0 0.14285715% reduce copy (9 of 21 at 0.00 MB/s)
2007-04-02 15:47:06,981 INFO org.apache.hadoop.mapred.TaskTracker:
task_0001_r_03_0 0.14285715% reduce copy (9 of 21 at 0.00 MB/s)
Pedro Guedes wrote:
Hi hadooping people
? If it's a problem with the wiki i'm very much willing
to help fix it...
Thanks in advance,
Pedro Guedes
Hi...
I was confused by the file://... changed it to (the obvious)
localhost:some_port and it works now... sorry for the silly questions,
and keep up the good work!
Pedro
Richard Yang wrote:
-Original Message-
From: Pedro Guedes [mailto:[EMAIL PROTECTED]
Sent: Wed 3/28/2007 8:04 AM