Hi Vangelis, Maybe your interval time is very little. That is caused fething every depth. Can you share nutch-site.xml and url's f coloumn fields and values.
Talat 11 May 2014 02:30 tarihinde "Vangelis karv" <[email protected]> yazdı: > Hi everyone! > > Let's say we start a crawl with depth 5 and topN 500 and www.something.com, > with domain(www.something.com) and regex urlfilters. > I have noticed that the url: www.something.com is fetched, parsed and > updated in every depth. Why is that happening? > In my opinion the particular url should be fetched and parsed only in the > 1st depth and updated in every depth. > > Thank you in advance, > Vangelis >

