I'm also thinking about implementing an automated workflow of
fetchlist->crawl->updateDb->index. Although my project may not require
NDSF because it only concerns about deep crawling of 100,000 sites, an
appropriate workflow is still needed to automatically take care of
failed urls, newly-added urls, daily update, etc. Appreciate it if
somebody can share experience on design of the workflow.
The nutch intranet crawler (or site-specific crawler, which I prefer to
call) is an automated process, but it's designed to conveniently deal
with just a handful of sites. With a larger number of selected sites, I
expect a modified version is needed. One modification I can think of is
to create a lookup table in the urlfilter object for domains to be
crawled and their corresponding regular expressions. The goal is to
avoid entering 100,000 regex in the craw-urlfilter.xml and checking ALL
these regex for each URL. Any comment?
thanks,
-AJ
Jay Lorenzo wrote:
Thanks, that's good information - it sounds like I need to take a closer
look at index deployment to see what the best solution is for automating
index management.
The initial email was more about understanding what the envisioned workflow
would for automating the creation of those indexes in a NDFS system, meaning
what
choices are available for automating the workflow of
fetchlist->crawl->updateDb->index
part of the equation when you have a node hosting a webdb, and a number of
nodes
crawling and indexing.
If I use a message based system, I assume I would create new fetchlists at a
given
locations of the NDFS, and message the fetchers where to find the
fetchlists. Once crawled,
I need to then update the webdb with the links discovered during the crawl.
Maybe this is too complex of a solution, but my sense is that map-reduce
systems still need a way
to manage the workflow/control that needs to occur if you want to create
pipelines that
generate indexes.
Thanks,
Jay Lorenzo
On 8/31/05, Doug Cutting <[EMAIL PROTECTED]> wrote:
I assume that in most NDFS-based configurations the production search
system will not run out of NDFS. Rather, indexes will be created
offline for a deployment (i.e., merging things to create an index
peractually
search node), then copied out of NDFS to the local filesystem on a
production search node and placed in production. This can be done
incrementally, where new indexes are deployed without re-deploying old
indexes. In this scenario, new indexes are rotated in replacing old
indexes, and the .del file for every index is updated, to reflect
deduping. There is no code yet which implements this.
Is this what you were asking?
Doug
Jay Lorenzo wrote:
I'm pretty new to nutch, but in reading through the mail lists and other
papers, I don't think I've really seen any discussion on using ndfs with
respect to automating end to end workflow for data that is going to be
searched (fetch->index->merge->search).
The few crawler designs I'm familiar with typically have spiders
(fetchers) and
indexers on the same box. Once pages are crawled and indexed the indexes
are pipelined to merge/query boxes to complete the workflow.
When I look at the nutch design and ndfs, I'm assuming the design intent
for 'pure ndfs' workflow is for the webdb to generate segments on a ndfs
partition, and once the updating of the webdb is completed, the segments
are processed 'on-disk' by the subsequent
fetcher/index/merge/query mechanisms. Is this a correct assumption?
Automating this kind of continuous workflow usually is dependent on the
implementation of some kind of control mechanism to assure that the
correct sequence of operations is performed.
Are there any recommendations on the best way to automate this
workflow when using ndfs? I've prototyped a continuous workflow system
using a traditional pipeline model with per stage work queues, and I see
how that could be applied to a clustered filesystem like ndfs, but I'm
curious to hear what the design intent or best practice is envisioned
for automating ndfs based implementations.
Thanks,
Jay
--
AJ (Anjun) Chen, Ph.D.
Canova Bioconsulting
Marketing * BD * Software Development
748 Matadero Ave., Palo Alto, CA 94306, USA
Cell 650-283-4091, [EMAIL PROTECTED]
---------------------------------------------------