Thanks! Indeed! It solved the issue!
Benjamin From: chethan <[email protected]> To: "[email protected]" <[email protected]>, Date: 30/04/2013 04:00 PM Subject: Re: Nutch 1.6 on Windows The version of Hadoop that Nutch uses in 1.6 causes this issue on Windows. This is resolved by changing the Hadoop dependency to *0.20.2* in $NUTCH_HOME/ivy/ivy.xml: <dependency org="org.apache.hadoop" name="hadoop-core" rev="0.20.2" conf="*->default"> <exclude org="hsqldb" name="hsqldb" /> <exclude org="net.sf.kosmosfs" name="kfs" /> <exclude org="net.java.dev.jets3t" name="jets3t" /> <exclude org="org.eclipse.jdt" name="core" /> <exclude org="org.mortbay.jetty" name="jsp-*" /> <exclude org="ant" name="ant" /> </dependency> This is what worked for me, although there could be a way of fixing this without rolling hadoop back to 0.20.2 that I'm unaware of. HTH Thanks Chethan On Tue, Apr 30, 2013 at 6:15 PM, Benjamin Sznajder <[email protected]>wrote: > > Hi, > > What are the needed software for running Nutch 1.6 on Windows? > > - I downloaded Nutch.1.6 > - I am using JDK 1.7 from IBM > - I installed Cygwin and am running Nutch from Cygwin. > > However, when launching the basic script, I am getting the following error. > > Are there specific steps that are necessary for running on Windows? > > Is Nutch supported to be run on Windows at all , i.e. not just for > development, but also in production? > > Best regards > Benjamin > > benjams@BENJAMS-TP /cygdrive/c/apache-nutch-1.6 > $ bin/nutch inject crawl/crawldb C:/temp/urls > cygpath: can't convert empty path > Injector: starting at 2013-04-30 15:42:09 > Injector: crawlDb: crawl/crawldb > Injector: urlDir: C:/temp/urls > Injector: Converting injected urls to crawl db entries. > Injector: java.io.IOException: Failed to set permissions of path: \tmp > \hadoop-benjams\mapred\staging\benjams-422849630\.staging to 0700 > at org.apache.hadoop.fs.FileUtil.checkReturnValue > (FileUtil.java:689) > at org.apache.hadoop.fs.FileUtil.setPermission(FileUtil.java:662) > at org.apache.hadoop.fs.RawLocalFileSystem.setPermission > (RawLocalFileSystem.java:509) > at org.apache.hadoop.fs.RawLocalFileSystem.mkdirs > (RawLocalFileSystem.java:344) > at org.apache.hadoop.fs.FilterFileSystem.mkdirs > (FilterFileSystem.java:189) > at org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir > (JobSubmissionFiles.java:116) > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:918) > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:912) > at java.security.AccessController.doPrivileged > (AccessController.java:314) > at javax.security.auth.Subject.doAs(Subject.java:572) > at org.apache.hadoop.security.UserGroupInformation.doAs > (UserGroupInformation.java:1149) > at org.apache.hadoop.mapred.JobClient.submitJobInternal > (JobClient.java:912) > at org.apache.hadoop.mapred.JobClient.submitJob (JobClient.java:886) > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1323) > at org.apache.nutch.crawl.Injector.inject(Injector.java:281) > at org.apache.nutch.crawl.Injector.run(Injector.java:318) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.nutch.crawl.Injector.main(Injector.java:308) > > >

