Hi, one way of fetching a file from disk is to configure the directory as local_url. In addition you need an index document, which "links" to all of your local files. So say local_urls: http://www.nonexistantserver.com/=/place/where/the/documents are The index document has to list your files like: http://www.nonexistantserver.com/dok1 http://www.nonexistantserver.com/dok2 http://www.nonexistantserver.com/dok3 http://www.nonexistantserver.com/dok4 http://www.nonexistantserver.com/dok5 set the start url to http://www.nonexistantserver.com/index_document.html That's it. To get a useful index of your XML documents, you may have to implement an external parser for your these files, but that's more complex :-( Hope this helps Cheers Stefan -----Urspr�ngliche Nachricht----- Von: Torbj�rn Gylleus [mailto:[EMAIL PROTECTED]] Gesendet: Freitag, 4. Mai 2001 16:14 An: [EMAIL PROTECTED] Betreff: [htdig] XML-import Hi everyone, I have been testing the crawler function of htdig and it seem verry nice. I also need to make indexes of structured data without fetching it by http, such as plain text files and XML data with url and description. Is this possible ? -T _______________________________________________ htdig-general mailing list <[EMAIL PROTECTED]> To unsubscribe, send a message to <[EMAIL PROTECTED]> with a subject of unsubscribe FAQ: http://htdig.sourceforge.net/FAQ.html _______________________________________________ htdig-general mailing list <[EMAIL PROTECTED]> To unsubscribe, send a message to <[EMAIL PROTECTED]> with a subject of unsubscribe FAQ: http://htdig.sourceforge.net/FAQ.html

