On Tue, 19 Jul 2005, Manuel Lemos wrote: > > We expect to produce a decently detailed refactoring document next week > > and create a 4.0 CVS branch then. > > Great. I hope that will allow us to do things like making Htdig crawl > individual pages and only update their entries in the index. That is > what miss most in the current HTDig version. > > I make htdig crawl the static version of my site every day, but that is > not very efficient and often it is too late. > > I can keep track of all pages that change and need to reindexed, but it > is odd to make Htdig crawl the hole site just because a few pages > changed. I would be more satisfied if I could just tell htdig once an > hour to reindex a limited list of pages that changed.
This should be exactly what Chris Murtagh's command does: echo URL_list.txt | htdig -m foo -c conf_file.conf - -m suppresses the addition of the full list of URLs in the db.docs to the 'to be requested queue' in the spider. Please reply back if this is not addressing your desires... Thanks -- Neal Richter Knowledgebase Developer RightNow Technologies, Inc. Customer Service for Every Web Site Office: 406-522-1485 ------------------------------------------------------- SF.Net email is sponsored by: Discover Easy Linux Migration Strategies from IBM. Find simple to follow Roadmaps, straightforward articles, informative Webcasts and more! Get everything you need to get up to speed, fast. http://ads.osdn.com/?ad_id=7477&alloc_id=16492&op=click _______________________________________________ ht://Dig Developer mailing list: [email protected] List information (subscribe/unsubscribe, etc.) https://lists.sourceforge.net/lists/listinfo/htdig-dev
