Group related sites together and use separate crawldb, segment directories.
On Fri, Nov 16, 2012 at 9:40 AM, Joe Zhang <[email protected]> wrote: > So how exactly do I set up different nutch instances then? > > On Thu, Nov 15, 2012 at 7:52 PM, Lewis John Mcgibbney < > [email protected]> wrote: > > > Hi Joe, > > > > In all honesty, it might sound slightly optimistic, it may also depend > > upon the size and calibre of the different sites/domains but if you > > are attempting a depth first, domain specific crawl, then maybe > > separate Nutch instances will be your friend... > > > > Lewis > > > > > > On Thu, Nov 15, 2012 at 11:53 PM, Joe Zhang <[email protected]> > wrote: > > > well, these are all details. The bigger question is, how to seperate > the > > > crawling policy of site A from that of site B? > > > > > > On Thu, Nov 15, 2012 at 7:41 AM, Sourajit Basak < > > [email protected]>wrote: > > > > > >> You probably need to customize parse-metatags plugin. > > >> > > >> I think you go ahead and include all possible metatags. And take care > of > > >> missing metatags in solr. > > >> > > >> On Thu, Nov 15, 2012 at 12:22 AM, Joe Zhang <[email protected]> > > wrote: > > >> > > >> > I understand conf/regex-urlfilter.txt; I can put domain names into > the > > >> URL > > >> > patterns. > > >> > > > >> > But what about meta tags? What if I want to parse out different meta > > tags > > >> > for different sites? > > >> > > > >> > On Wed, Nov 14, 2012 at 1:33 AM, Sourajit Basak < > > >> [email protected] > > >> > >wrote: > > >> > > > >> > > 1) For parsing & indexing customized meta tags enable & configure > > >> plugin > > >> > > "parse-metatags" > > >> > > > > >> > > 2) There are several filters of url, like regex based. For regex, > > the > > >> > > patterns are specified via conf/regex-urlfilter.txt > > >> > > > > >> > > On Wed, Nov 14, 2012 at 1:33 PM, Tejas Patil < > > [email protected] > > >> > > >wrote: > > >> > > > > >> > > > While defining url patterns, have the domain name in it so that > > you > > >> get > > >> > > > site/domain specific rules. I don't know about configuring meta > > tags. > > >> > > > > > >> > > > Thanks, > > >> > > > Tejas > > >> > > > > > >> > > > > > >> > > > On Tue, Nov 13, 2012 at 11:34 PM, Joe Zhang < > [email protected] > > > > > >> > > wrote: > > >> > > > > > >> > > > > How to enforce site-specific crawling policies, i.e, different > > URL > > >> > > > > patterns, meta tags, etc. for different websites to be > crawled? > > I > > >> got > > >> > > the > > >> > > > > sense that multiple instances of nutch are needed? Is it > > correct? > > >> If > > >> > > yes, > > >> > > > > how? > > >> > > > > > > >> > > > > > >> > > > > >> > > > >> > > > > > > > > -- > > Lewis > > >

