Hi Scott,
yes this makes a sense.
I would also create a temp web db create the segment, crawl the segment.
If you don't want to add the pages down of the new urls than just
index the segment and add this segment to the other searchable
segments, do not update the db.
In general if you inject a set of urls to a webdb and create new
segment the segment should only contains the new urls and pages that
are older than 30 days and fetched anyway.
Greetings,
Stefan
Am 08.02.2006 um 14:56 schrieb Scott Owens:
Hi All,
I wanted to check in to see if anyone has found an answer for this
issue. I am injecting new URLs on a daily basis, and only need to
fetch/index those new one's, but obviously need to maintain a complete
webdb.
One thing I was thinking was to use a temporary webdb for the initial
injection, then updating (updatedb) my primary webdb after the fetch
or indexing.
# prepare dirs and inject urls
rm -rf $db/*
$nutch admin -local $db -create
$nutch inject -local $db -urlfile $urlFile
echo -e "\nGenerating next segment to fetch"
$nutch generate -local $db $segmentdir $fetchLimit
s=`ls -d $segmentdir/* | tail -1`
echo -e "\nFetching next segment"
$nutch fetch $s
echo -e "\nUpdating web database"
$nutch updatedb $dbmain $s
echo -e "\nAnalyzing links"
$nutch analyze $dbmain 5
OR after the segment is indexed -- as the above method wouldn't allow
a depth greather than 1?
# prepare dirs and inject urls
rm -rf $db/*
$nutch admin -local $db -create
$nutch inject -local $db -urlfile $urlFile
for i in `seq $depth`
do
echo -e "\nGenerating next segment to fetch"
$nutch generate -local $db $segmentdir $fetchLimit
s=`ls -d $segmentdir/* | tail -1`
echo -e "\nFetching next segment"
$nutch fetch $s
echo -e "\nUpdating web database"
$nutch updatedb $db $s
echo -e "\nAnalyzing links"
$nutch analyze $db 5
done
echo -e "\nFetch done"
echo "Indexing segments"
for s in `ls -1d $segmentdir/*`
do
$nutch index $s
done
echo -e "\nUpdating web database"
$nutch updatedb $dbmain $s
OR maybe I have no idea what I'm talking about : ) - I'm not a
developer, just trying to figure things out.
If anyone has experience with this and some advice I'm all ears.
thanks!
Scott
On 11/10/05, Dean Elwood <[EMAIL PROTECTED]> wrote:
Hi Lawrence,
I'm stuck in the same position. I haven't yet examined the "merge"
function,
which might shed some light on it.
Have you managed to discover anything so far?
You can use the regular expression bases url filter. Than only
urls that
match the pattern will be added to a fetch list.<<
Hi Stefan. Getting the new URL's to crawl is the easy part ;-)
The trick, and the question, is how you add that to an existing
database,
and then re-index, without doing a full re-crawl?
Thanks,
Dean
----- Original Message -----
From: "Lawrence Pitcher" <[EMAIL PROTECTED]>
To: <[email protected]>
Sent: Thursday, November 10, 2005 5:05 PM
Subject: How to add only new urls to DB
Hi,
Thanks to all for the best search solution available.
I have installed the software, indexed 15,000 websites and tested
the search
and it works great!
If I want to add only two more websites, so I made a "newurls.txt"
file,
then injected it to WebDB "bin/nutch inject db/ -urlfile
newurls.txt", then
generated a new segment "bin/nutch generate db/ segments/", I
then checked
for the new sement name in the directory "/segments'
Took that new segment name and placed it in the fetch command "bin/
nutch
fetch segments/20051110103316/"
However it appears to re-fetch all 15,000 webpages along with the
newurls.txt webpages.
Can I not just index only the new and then Update the DB.
Sorry for such a lame question but I have just started.
Many thanks to all.
Lawrence
---------------------------------------------------------------
company: http://www.media-style.com
forum: http://www.text-mining.org
blog: http://www.find23.net