Hi,

It seems, that I'm not the only having "Target /tmp/.../map_ynynnj.out
already exists" error. Teruhiko Kurosaka is experiencing the same
trouble. I advised him to use this version of nutch
http://68.178.249.66/nutch-admin/nutch-0.8-dev_guiBundle_05_02_06.tar.gz
and it works, but the current version is causing this error. Any ideas?

I have submitted bugreport, but still no response to it.
http://issues.apache.org/jira/browse/NUTCH-266

Btw, in the version with gui there's a problem with temporary hadoop
map-reduce files. They are constantly created, which leads to the high
space usage. Is it solved in the current build?

> Hello,
> I am still unable to run "nutch crawl", which terminates with "Job
> Failed!"
> IO Exception.
> In an attempt to get more info, I increased the logging level and ran 
> "nutch crawl" again.  Now it is clear that nutch is failing in renaming
> a file.

> 2006-06-16 17:04:05,932 INFO  mapred.LocalJobRunner
> (LocalJobRunner.java:progress(140)) -
> C:/opt/nutch-060614/test/index/segments/20060616170358/crawl_parse/part-
> 00000:0+62
> 2006-06-16 17:04:05,948 WARN  mapred.LocalJobRunner
> (LocalJobRunner.java:run(119)) - job_4wsxze
> java.io.IOException: Couldn't rename
> /tmp/hadoop/mapred/local/map_5n5aid/part-0.out
>         at
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:102)
> Exception in thread "main" java.io.IOException: Job failed!
>         at
> org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:342)
>         at org.apache.nutch.crawl.CrawlDb.update(CrawlDb.java:55)
>         at org.apache.nutch.crawl.Crawl.main(Crawl.java:114)

> I am wondering what LocalJobRunner is trying to accomplish.  Anybody?


> In addition to this fatal exception, I've seen many occurances of this
> exception:
> 2006-06-16 17:04:05,854 INFO  conf.Configuration
> (Configuration.java:loadResource(397)) - parsing
> file:/C:/opt/nutch-060614/conf/hadoop-site.xml
> 2006-06-16 17:04:05,870 DEBUG conf.Configuration
> (Configuration.java:<init>(67))
>  - java.io.IOException: config()
>         at
> org.apache.hadoop.conf.Configuration.<init>(Configuration.java:67)
>         at org.apache.hadoop.mapred.JobConf.<init>(JobConf.java:115)
>         at
> org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:6
> 1)
>         at
> org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:18
> 1)
>         at
> org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:277)
>         at
> org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:312)
>         at org.apache.nutch.crawl.CrawlDb.update(CrawlDb.java:55)
>         at org.apache.nutch.crawl.Crawl.main(Crawl.java:114)

>  
> Is this the cause of the fatal exception?


> I am not intend to run hadoop at all, so this hadoop-site.xlm is empty.
> It just has
> <configuration>
> </configuration>



> Somebody told me of the following binary package, and this one carwls
> fine.
> http://68.178.249.66/nutch-admin/nutch-0.8-dev_guiBundle_05_02_06.tar.gz
> So, some code chages that were introduced within the last few weeks must
> be causing this problem.

-- 
Best regards,
 Eugen                            mailto:[EMAIL PROTECTED]



_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to