Hello,

I would like to index many files using a mysql database. Before a file 
should be indexed I would like to check the content with a perl script, 
which also decides if the file is worth to be indexed.

Before a file will be downloaded I would like to check with a perl code 
if the link seams to be interesting and make a decision according to the 
link name.

The Web Interface to access the crawlers database should be in PHP. The 
layout I would like to make like google. If it is possible to implement 
own ranking modules it would be fine.

Currently I use LWP::RobotUA and write the interesting files completely 
in a mysql database and search the database with LIKE '%$WORD%' which is 
very slow.

Now there are multiple systems available:
  DBIx::KwIndex, DBIx::Fulltext, DBIx::TextIndex
  Glimpse, htdig, SIWSH-E, Isearch, WordIndex

Could somebody recommend on of this systems for this needs.

Many thanks

Matthias Jaekle






_______________________________________________
Robots mailing list
[EMAIL PROTECTED]
http://www.mccmedia.com/mailman/listinfo/robots

Reply via email to