Hi Roannel, we go try it Thanks,
Paul 2015-11-19 10:08 GMT-05:00 Roannel Fernández Hernández <[email protected]>: > Hi Paul > > Include in your plugin.includes property the scoring-depth plugin and set > the value 1 in your property scoring.depth.max. > > See: https://issues.apache.org/jira/browse/NUTCH-1331 for more > information. > > Regards. > > ----- Mensaje original ----- > > De: "Paul Escobar" <[email protected]> > > Para: [email protected] > > Enviados: Miércoles, 18 de Noviembre 2015 22:33:50 > > Asunto: Re: [MASSMAIL]Crawling focused only over seed file > > > > Hi Roannel, the new URLs aren't from other domains, they are in the same > > domain, we want updatedb command avoid the update crawldb with new url > from > > the same site. > > > > Thanks, > > > > Paul > > > > 2015-11-18 21:57 GMT-05:00 Andrés Rincón Pacheco <[email protected]>: > > > > > Hi Roannel, > > > > > > I had the parameter configured previously but this not solved the > problem. > > > How I can avoid add any newly discovered URLs during fetch process? I > want > > > that nutch process only urls of seed file. > > > > > > Thanks. > > > > > > > > > 2015-11-18 9:22 GMT-05:00 Roannel Fernández Hernández <[email protected] > >: > > > > > > > Hi Andrés, > > > > > > > > Change in your nutch-site.xml the property db.ignore.external.links > to > > > > true. > > > > > > > > Regards > > > > > > > > ----- Mensaje original ----- > > > > > De: "Andrés Rincón Pacheco" <[email protected]> > > > > > Para: [email protected] > > > > > Enviados: Sábado, 14 de Noviembre 2015 19:51:54 > > > > > Asunto: [MASSMAIL]Crawling focused only over seed file > > > > > > > > > > Hi, > > > > > > > > > > I need execute nutch focus over seed file, no more urls added in > every > > > > > cycle. > > > > > > > > > > I am executing nutch with the following scenarios: > > > > > > > > > > 1. Invoking crawl script without updatedb job: The time of > execution > > > for > > > > > every cycle is 15 minutes, but > > > > > in every cycle the urls processing are the same. The total time > for > > > > nutch > > > > > execution is around 16 hours. > > > > > Because the urls in every cycle are the same? > > > > > > > > > > 2. Crawling normal (using updateddb): if I am using updatedb job, > how > > > can > > > > > nutch make fetch only urls of seed file without add new urls to > > > crawldb? > > > > > > > > > > I am trying execute nutch using updatedb job with -noAdditions, so > that > > > > it > > > > > serves this option? I was reading the nutch wiki but is not clear > the > > > > > performance of > > > > > -noAdditions option > > > > > > > > > > Conditions for every case: the configuration used for proccessing > is > > > 360 > > > > > urls in every cycle. The seed file contains around 25000 urls. > > > > > (limit parameter in crawl bash script is 25000 and sizeFetchlist is > > > 360). > > > > > > > > > > Thanks, > > > > > > > > > > Andres > > > > > > > > > Noviembre 13-14: Final Caribeña 2015 del Concurso de Programación > > > ACM-ICPC > > > > https://icpc.baylor.edu/regionals/finder/cf-2015 > > > > > > > > > > > > > > > -- > > Paul Escobar Mossos > > skype: paulescom > > telefono: +57 1 3006815404 > > > Noviembre 13-14: Final Caribeña 2015 del Concurso de Programación ACM-ICPC > https://icpc.baylor.edu/regionals/finder/cf-2015 > -- Paul Escobar Mossos skype: paulescom telefono: +57 1 3006815404

