I'm getting the error below. When I look at the rootUrlFile value, it
seems as though it is trying to read a file named "urls.txt -dir
crawled" rather than recognizing the -dir parameter. Any ideas? This is
running on Solaris 9, if that makes any difference.

If I merely run "bin/nutch crawl urls.txt" this problem doesn't occur.

Thanks!


bash-2.05# bin/nutch crawl urls.txt -dir crawled
060414 155951 parsing
file:/export/home/www/virtual/wiki/doc_root/nutch-0.7.2/conf/nutch-default.xml
060414 155951 parsing
file:/export/home/www/virtual/wiki/doc_root/nutch-0.7.2/conf/crawl-tool.xml
060414 155951 parsing
file:/export/home/www/virtual/wiki/doc_root/nutch-0.7.2/conf/nutch-site.xml
060414 155951 No FS indicated, using default:local
060414 155951 crawl started in: crawl-20060414155951
060414 155951 rootUrlFile = urls.txt -dir crawled
060414 155951 threads = 10
060414 155951 depth = 5
060414 155952 Created webdb at
LocalFS,/export/home/www/virtual/wiki/doc_root/nutch-0.7.2/crawl-20060414155951/db
Exception in thread "main" java.io.FileNotFoundException: urls.txt -dir
crawled (No such file or directory)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.<init>(FileInputStream.java:106)
at java.io.FileReader.<init>(FileReader.java:55)
at org.apache.nutch.db.WebDBInjector.injectURLFile(WebDBInjector.java:372)
at org.apache.nutch.db.WebDBInjector.main(WebDBInjector.java:535)
at org.apache.nutch.tools.CrawlTool.main(CrawlTool.java:134)



-------------------------------------------------------
This SF.Net email is sponsored by xPML, a groundbreaking scripting language
that extends applications into web and mobile media. Attend the live webcast
and join the prime developer group breaking into this new coding territory!
http://sel.as-us.falkag.net/sel?cmd=lnk&kid=110944&bid=241720&dat=121642
_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to