Hello, I've just released a modified version of nutch071 and tomcat50 running off a CDROM or local harddrive cross-platform:
http://sf.net/projects/vicaya My ambitions are not 'the whole web' but a small and static collection of pages. I intend to allow users to use nutch offline with the occasional online content and index update (RSS, webstart, and/or Subversion). Please let me know if such questions are out of scope. I have found that reading the segments on CDROM is the biggest performance bottleneck. However, I do not want to require that the user copies the entire segments directory to disk. Is it possible to separate some data - such as the reverse index from the other fields? Would this require a change to Lucene or Nutch's source code? I am considering importing content and index segments into an SVN repository so that users may receive periodic updates. Will the segments directory lend itself well to SVN patches? I have experimented mostly with intranet search, but I've noticed that whole web search creates dated indices. Might it be a matter of adding new crawl segments since the last update? Thanks, Alex -- Those who can make you believe absurdities can make you commit atrocities -- François Marie Arouet (Voltaire) http://cph.blogsome.com http://genaud.org/alex/key.asc -- CCC7 D19D D107 F079 2F3D BF97 8443 DB5A 6DB8 9CE1
