Hi, It seems, that I'm not the only having "Target /tmp/.../map_ynynnj.out already exists" error. Teruhiko Kurosaka is experiencing the same trouble. I advised him to use this version of nutch http://68.178.249.66/nutch-admin/nutch-0.8-dev_guiBundle_05_02_06.tar.gz and it works, but the current version is causing this error. Any ideas?
I have submitted bugreport, but still no response to it. http://issues.apache.org/jira/browse/NUTCH-266 Btw, in the version with gui there's a problem with temporary hadoop map-reduce files. They are constantly created, which leads to the high space usage. Is it solved in the current build? > Hello, > I am still unable to run "nutch crawl", which terminates with "Job > Failed!" > IO Exception. > In an attempt to get more info, I increased the logging level and ran > "nutch crawl" again. Now it is clear that nutch is failing in renaming > a file. > 2006-06-16 17:04:05,932 INFO mapred.LocalJobRunner > (LocalJobRunner.java:progress(140)) - > C:/opt/nutch-060614/test/index/segments/20060616170358/crawl_parse/part- > 00000:0+62 > 2006-06-16 17:04:05,948 WARN mapred.LocalJobRunner > (LocalJobRunner.java:run(119)) - job_4wsxze > java.io.IOException: Couldn't rename > /tmp/hadoop/mapred/local/map_5n5aid/part-0.out > at > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:102) > Exception in thread "main" java.io.IOException: Job failed! > at > org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:342) > at org.apache.nutch.crawl.CrawlDb.update(CrawlDb.java:55) > at org.apache.nutch.crawl.Crawl.main(Crawl.java:114) > I am wondering what LocalJobRunner is trying to accomplish. Anybody? > In addition to this fatal exception, I've seen many occurances of this > exception: > 2006-06-16 17:04:05,854 INFO conf.Configuration > (Configuration.java:loadResource(397)) - parsing > file:/C:/opt/nutch-060614/conf/hadoop-site.xml > 2006-06-16 17:04:05,870 DEBUG conf.Configuration > (Configuration.java:<init>(67)) > - java.io.IOException: config() > at > org.apache.hadoop.conf.Configuration.<init>(Configuration.java:67) > at org.apache.hadoop.mapred.JobConf.<init>(JobConf.java:115) > at > org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:6 > 1) > at > org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:18 > 1) > at > org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:277) > at > org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:312) > at org.apache.nutch.crawl.CrawlDb.update(CrawlDb.java:55) > at org.apache.nutch.crawl.Crawl.main(Crawl.java:114) > > Is this the cause of the fatal exception? > I am not intend to run hadoop at all, so this hadoop-site.xlm is empty. > It just has > <configuration> > </configuration> > Somebody told me of the following binary package, and this one carwls > fine. > http://68.178.249.66/nutch-admin/nutch-0.8-dev_guiBundle_05_02_06.tar.gz > So, some code chages that were introduced within the last few weeks must > be causing this problem. -- Best regards, Eugen mailto:[EMAIL PROTECTED] _______________________________________________ Nutch-general mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/nutch-general
