Hi Roannel,

I had the parameter configured previously but this not solved the problem.
How I can avoid add any newly discovered URLs during fetch process? I want
that nutch process only urls of seed file.

Thanks.


2015-11-18 9:22 GMT-05:00 Roannel Fernández Hernández <[email protected]>:

> Hi Andrés,
>
> Change in your nutch-site.xml the property db.ignore.external.links to
> true.
>
> Regards
>
> ----- Mensaje original -----
> > De: "Andrés Rincón Pacheco" <[email protected]>
> > Para: [email protected]
> > Enviados: Sábado, 14 de Noviembre 2015 19:51:54
> > Asunto: [MASSMAIL]Crawling focused only over seed file
> >
> > Hi,
> >
> > I need execute nutch focus over seed file, no more urls added in every
> > cycle.
> >
> > I am executing nutch with the following scenarios:
> >
> > 1. Invoking crawl script without updatedb job:  The time of execution for
> > every cycle is 15 minutes, but
> > in every cycle the urls processing are the same.  The total time for
> nutch
> > execution is around 16 hours.
> > Because the urls in every cycle are the same?
> >
> > 2. Crawling normal (using updateddb): if I am using updatedb job, how can
> > nutch make fetch only urls of seed file without add new urls to crawldb?
> >
> > I am trying execute nutch using updatedb job with -noAdditions, so that
> it
> > serves this option?  I was reading the nutch wiki but is not clear the
> > performance of
> > -noAdditions option
> >
> > Conditions for every case: the configuration used for proccessing is 360
> > urls in every cycle.  The seed file contains around 25000 urls.
> > (limit parameter in crawl bash script is 25000 and sizeFetchlist is 360).
> >
> > Thanks,
> >
> > Andres
> >
> Noviembre 13-14: Final Caribeña 2015 del Concurso de Programación ACM-ICPC
> https://icpc.baylor.edu/regionals/finder/cf-2015
>

Reply via email to