Hi,

I have manual script for my first crawl, anyone can explain this command
step by step:

*Initialize the crawldb*
bin/nutch inject urls/
*Generate URLs from crawldb*
bin/nutch generate -topN 80
*Fetch generated URLs*
bin/nutch fetch -all
*Parse fetched URLs*
bin/nutch parse -all
*Update database from parsed URLs*
bin/nutch updatedb -all
*Index parsed URLs*
bin/nutch index -all

anyone can help me  how re-crawling script.



Thanks


Regard's

Muchlis

Reply via email to