Hi Paul

Include in your plugin.includes property the scoring-depth plugin and set the 
value 1 in your property scoring.depth.max.

See: https://issues.apache.org/jira/browse/NUTCH-1331 for more information.

Regards.

----- Mensaje original -----
> De: "Paul Escobar" <[email protected]>
> Para: [email protected]
> Enviados: Miércoles, 18 de Noviembre 2015 22:33:50
> Asunto: Re: [MASSMAIL]Crawling focused only over seed file
> 
> Hi Roannel, the new URLs aren't from other domains, they are in the same
> domain, we want updatedb command avoid the update crawldb with new url from
> the same site.
> 
> Thanks,
> 
> Paul
> 
> 2015-11-18 21:57 GMT-05:00 Andrés Rincón Pacheco <[email protected]>:
> 
> > Hi Roannel,
> >
> > I had the parameter configured previously but this not solved the problem.
> > How I can avoid add any newly discovered URLs during fetch process? I want
> > that nutch process only urls of seed file.
> >
> > Thanks.
> >
> >
> > 2015-11-18 9:22 GMT-05:00 Roannel Fernández Hernández <[email protected]>:
> >
> > > Hi Andrés,
> > >
> > > Change in your nutch-site.xml the property db.ignore.external.links to
> > > true.
> > >
> > > Regards
> > >
> > > ----- Mensaje original -----
> > > > De: "Andrés Rincón Pacheco" <[email protected]>
> > > > Para: [email protected]
> > > > Enviados: Sábado, 14 de Noviembre 2015 19:51:54
> > > > Asunto: [MASSMAIL]Crawling focused only over seed file
> > > >
> > > > Hi,
> > > >
> > > > I need execute nutch focus over seed file, no more urls added in every
> > > > cycle.
> > > >
> > > > I am executing nutch with the following scenarios:
> > > >
> > > > 1. Invoking crawl script without updatedb job:  The time of execution
> > for
> > > > every cycle is 15 minutes, but
> > > > in every cycle the urls processing are the same.  The total time for
> > > nutch
> > > > execution is around 16 hours.
> > > > Because the urls in every cycle are the same?
> > > >
> > > > 2. Crawling normal (using updateddb): if I am using updatedb job, how
> > can
> > > > nutch make fetch only urls of seed file without add new urls to
> > crawldb?
> > > >
> > > > I am trying execute nutch using updatedb job with -noAdditions, so that
> > > it
> > > > serves this option?  I was reading the nutch wiki but is not clear the
> > > > performance of
> > > > -noAdditions option
> > > >
> > > > Conditions for every case: the configuration used for proccessing is
> > 360
> > > > urls in every cycle.  The seed file contains around 25000 urls.
> > > > (limit parameter in crawl bash script is 25000 and sizeFetchlist is
> > 360).
> > > >
> > > > Thanks,
> > > >
> > > > Andres
> > > >
> > > Noviembre 13-14: Final Caribeña 2015 del Concurso de Programación
> > ACM-ICPC
> > > https://icpc.baylor.edu/regionals/finder/cf-2015
> > >
> >
> 
> 
> 
> --
> Paul Escobar Mossos
> skype: paulescom
> telefono: +57 1 3006815404
> 
Noviembre 13-14: Final Caribeña 2015 del Concurso de Programación ACM-ICPC
https://icpc.baylor.edu/regionals/finder/cf-2015

Reply via email to