> in the Database all(let s say 24) pages are stored.
The Database Stored 24 "URLs". That is the one URL which is Indexed an
the 23 URLs which are on the linked on the page an Nutch must Index in
the next crawl.
Best regards from Germany
Michael
Nils Hoeller schrieb:
Hi,
i ve got following Problem.
When I crawl and index a Site
with for example depth 1, it
works perfectly for the WebDB which means,
in the Database all(let s say 24) pages are stored.
But when I look at the index Dir with Luke
I see only one page/doc (root page of crawl).
Now when I increase the depth of the crawl
to 2, I have about 400 pages in the
WebDB and the 24 in the Index.
So the Index seems to be made for depth-1 # of Pages?
Why is that so ? Is that a configuration problem ?
Thanks for your help
Nils
-------------------------------------------------------
SF.Net email is Sponsored by the Better Software Conference & EXPO
September 19-22, 2005 * San Francisco, CA * Development Lifecycle Practices
Agile & Plan-Driven Development * Managing Projects & Teams * Testing & QA
Security * Process Improvement & Measurement * http://www.sqe.com/bsce5sf
_______________________________________________
Nutch-developers mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-developers