Hi,

one way of fetching a file from disk is to configure the directory as
local_url. In addition you need an index document, which "links" to all of
your local files.

So say local_urls:
http://www.nonexistantserver.com/=/place/where/the/documents are

The index document has to list your files like:

http://www.nonexistantserver.com/dok1
http://www.nonexistantserver.com/dok2
http://www.nonexistantserver.com/dok3
http://www.nonexistantserver.com/dok4
http://www.nonexistantserver.com/dok5

set the start url to http://www.nonexistantserver.com/index_document.html

That's it. 

To get a useful index of your XML documents, you may have to implement an
external parser for your these files, but that's more complex :-(

Hope this helps

  Cheers
 
    Stefan





-----Urspr�ngliche Nachricht-----
Von: Torbj�rn Gylleus [mailto:[EMAIL PROTECTED]]
Gesendet: Freitag, 4. Mai 2001 16:14
An: [EMAIL PROTECTED]
Betreff: [htdig] XML-import


Hi everyone,

I have been testing the crawler function of htdig and it seem verry nice. I
also need to make indexes of structured data without fetching it by http,
such as plain text files and XML data with url and description.
Is this possible ?

-T


_______________________________________________
htdig-general mailing list <[EMAIL PROTECTED]>
To unsubscribe, send a message to
<[EMAIL PROTECTED]> with a subject of unsubscribe
FAQ: http://htdig.sourceforge.net/FAQ.html

_______________________________________________
htdig-general mailing list <[EMAIL PROTECTED]>
To unsubscribe, send a message to <[EMAIL PROTECTED]> with a 
subject of unsubscribe
FAQ: http://htdig.sourceforge.net/FAQ.html

Reply via email to