I have a 1.6 gigabyte database that is the result of 3 or 4 days worth of indexing 3 websites defined by the /usr/local/share/htdig/common/start.url file.
I used the /usr/local/bin/rundig file which seems to have worked. However, I did not pipe the output to a file : (
-rw-r--r-- 1 root wheel 1.6G Dec 23 20:39 db.docdb -rw-r--r-- 1 root wheel 2.0k Dec 23 18:55 db.docs.index -rw-r--r-- 1 root wheel 163M Dec 23 18:55 db.wordlist -rw-r--r-- 1 root wheel 105M Dec 23 18:55 db.words.db
The htsearch in cgi-bin loads up fine but I cannot find any words at all whatsoever, no matter what I search for.
I have now increased the size of the following defaults, but not yet re-run the rundig script because I know it is going to take days....
#max_head_length: 10000 max_head_length: 100000
#max_doc_size: 200000 max_doc_size: 2000000
but I seem to recall that the last time I attempted to index these sites, (.jsp driven) that helped -- maybe??.
The problem is that is going to take days to run and might not help.
So, my question is "WTF is in that 1.6 gig database and how do I know in advance what the current problem is so that I can fix it?"
Thanks,
Rick
------------------------------------------------------- This SF.net email is sponsored by: IBM Linux Tutorials. Become an expert in LINUX or just sharpen your skills. Sign up for IBM's Free Linux Tutorials. Learn everything from the bash shell to sys admin. Click now! http://ads.osdn.com/?ad_id=1278&alloc_id=3371&op=click _______________________________________________ ht://Dig general mailing list: <[EMAIL PROTECTED]> ht://Dig FAQ: http://htdig.sourceforge.net/FAQ.html List information (subscribe/unsubscribe, etc.) https://lists.sourceforge.net/lists/listinfo/htdig-general

