I prefer a suite of shell scripts and cron jobs. We simply generate many 
segments at once, have a cron job checking for available segments we can fetch 
and fetch them. If all are fetched, the segemnts are moved to a queue 
directory for updating the DB. Once the DB has been updated the generators are 
triggered and the whole circus repeats.


> I've done some searching on this, but haven't found any real solutions.  Is
> there an existing way to do a continuous crawl using Nutch?  I know I can
> use the bin/nutch crawl command, but that stops after a certain number of
> iterations.
> 
> Right now I'm working on a java class to do it, but I would assume it's a
> problem that's been solved already.  Unfortunately I can't seem to find any
> evidence of this.
> 
> Thanks.

Reply via email to