Hi Vangelis,

Maybe your interval time is very little. That is caused fething every
depth. Can you share nutch-site.xml and url's f coloumn fields and values.

Talat
11 May 2014 02:30 tarihinde "Vangelis karv" <[email protected]> yazdı:

> Hi everyone!
>
> Let's say we start a crawl with depth 5 and topN 500 and www.something.com,
> with domain(www.something.com) and regex urlfilters.
> I have noticed that the url: www.something.com is fetched, parsed and
> updated in every depth. Why is that happening?
> In my opinion the particular url should be fetched and parsed only in the
> 1st depth and updated in every depth.
>
> Thank you in advance,
> Vangelis
>

Reply via email to