Is there anything like this that would run on a Windows 98 or NT platform?

> Jim MacDiarmid, Senior Software Engineer
> PACEL Corp.
> 8870 Rixlew Lane
> Manassas, VA 20109
> (703) 257-4759
> FAX:  (703) 361-6706
> www.pacel.com
>
>
>
> -----Original Message-----
> From: Simon Wilkinson [SMTP:[EMAIL PROTECTED]]
> Sent: Sunday, January 14, 2001 4:37 PM
> To:   [EMAIL PROTECTED]
> Subject:      Re: Looking for a gatherer.
>
> > I am looking for a spider/gatherer with the following characteristics:
> >     * Enables the control of the crawling process by URL
> substring/regexp
> > and HTML context of the link.
> >     * Enables the control of the gathering (i.e. saving) processes by
> URL
> > substring/regexp, MIME type, other header information and ideally by
> some
> > predicates on the HTML source.
> >     * Some way to save page/document metadata, ideally in a database.
> >     * Freeware, shareware or otherwise inexpensive would be nice.
>
> You might like to take a look at Harvest-NG, which is free software.
> (http://webharvest.sourceforge.net/ng) It will allow all of what you
> detail above. It saves the metadata in a Perl DBM database - some work
> has been done, but not completed, on working with the DBI interface
> to a remote database. You may find that some knowledge of Perl is helpful
> in adapting it exactly to your needs (much use is made of Perl regular
> expressions in the pattern matching, for instance).
>
> Cheers,
>
> Simon.

Reply via email to