Cool.

Thanks & Regards,
Preetam

On Sun, Feb 8, 2015 at 11:16 AM, Mattmann, Chris A (3980) <
[email protected]> wrote:

> Thanks Preetam:
>
>
> >[..snip..]
> >Why would you want to?
> >
> >
> >
> >
> >Preetam : Just was curious to manually handle this if possible.
> >I was anticipating that once the db has been fetched and CrawlDB has all
> >the URLs crawled data to depth 2, the next run should not crawl the same
> >URLs again.
> >Is it that URLs fetched at depth 2 which are kept unfetched in the queue
> >(not deque'd since  it has crawled the threshold depth passed) due to the
> >depth value constraint and are hence fetched in the next run resulting in
> >increase in fetch size ?
>
> +1. Yep.
>
> Cheers,
> Chris
>
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> Chris Mattmann, Ph.D.
> Chief Architect
> Instrument Software and Science Data Systems Section (398)
> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> Office: 168-519, Mailstop: 168-527
> Email: [email protected]
> WWW:  http://sunset.usc.edu/~mattmann/
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> Adjunct Associate Professor, Computer Science Department
> University of Southern California, Los Angeles, CA 90089 USA
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>
>

Reply via email to