> Is it the old 2048 MB limit on filesize?  Linux supports longs
> for fseek in the like, is htdig limited to fseeking with int's?
>
> Is this a berkeley db limitation?

In some sense, yes, in some sense no. For one, keep in mind that it's 
not *Linux* per-se that has any problems with 2GB files as Linux on 
64-bit systems have no problems with this. (Though as I've pointed out 
in the past, FreeBSD and Solaris/x86 have no problem with it either.) 
IIRC, it's also more than just an x86-Linux kernel issue, but a bit more 
complicated (glibc is also involved).

The Berkeley DB code when compiled tries to figure out whether the OS 
has the 2GB limitation. You should look at the configure output from the 
db about "bigfile" and you may wish to try compiling it separately from 
ht://Dig with ./configure --enable-bigfile

I'm not sure if the configure flag will force it to try bigfile support 
if the configure script feels your OS doesn't properly support them.

> Does Htdig 3.2.X have the same limitation?

Sometimes, as above. Though it's harder to break the limit since the 
word database is never written as an ASCII file, always compressed, and 
the document database has the db.excerpts split out.

> Are there any guidelines (I checked the faq) for the number of url's
> that are reasonable to index?

Not really. Depends on how much of an excerpt you store, how big your 
URLs tend to be, whether you're running on a system with a 2GB 
limitation... Some users even back in the 3.1.0 days had installations 
of 10 million URLs on Solaris systems.

--
-Geoff Hutchison
Williams Students Online
http://wso.williams.edu/


_______________________________________________
htdig-general mailing list <[EMAIL PROTECTED]>
To unsubscribe, send a message to <[EMAIL PROTECTED]> with a 
subject of unsubscribe
FAQ: http://htdig.sourceforge.net/FAQ.html

Reply via email to