I tried to Index my  local file system according to the FAQ: 
http://wiki.apache.org/nutch/FAQ#head-c721b23b43b15885f5ea7d8da62c1c40a37878e6

But if I add the plugin into the nutch-site.xml file like this:

      <property>
        <name>plugin.includes</name>
       
<value>protocol-file|protocol-http|parse-(text|html)|index-basic|query-(basic|site|url)</value>
      </property>

There will be a Exception:

Injector: Converting injected urls to crawl db entries.
Exception in thread "main" java.io.IOException: Job failed!
        at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:357)
        at org.apache.nutch.crawl.Injector.inject(Injector.java:138)
        at org.apache.nutch.crawl.Crawl.main(Crawl.java:105)

Probably the reason is the bug described in:
 * http://issues.apache.org/jira/browse/NUTCH-384

As a hack I could use a (local) webserver to feed nutch with the files.
But maybe there is a better workaround to index from a local filesystem
with nutch 8.x?
Can you help me?

Additionally I have another question:
 * Is there a possibility to use a directory of the HDFS Filesystem as a
spool directory to index from?


Thanks

Christian Herta


-------------------------------------------------------------------------
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT & business topics through brief surveys - and earn cash
http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
_______________________________________________
Nutch-general mailing list
Nutch-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to