At 8:35 AM -0500 11/24/99, [EMAIL PROTECTED] wrote:
>I've got a rather large page (over 200,000 links to very simple pages with
>no links), that I am trying to index.  Whenever I run htdig against it, I
>eventually get the error "virtual memory exceeded in 'new'".
>
>I really don't want to split the page into lots of smaller ones because it's
>dynamically generated from a database so... is there any way I can index
>these files without getting that error?

Yes, but only if you increase the amount of swap space...

Here's the problem--in order to do the indexing, htdig must assemble 
a TODO list of links. Since indexing that page just added 200,000 
links to the TODO list, it needs to grab a bunch more memory. But 
that's the problem it mentioned--it ran out of memory (in this case 
swap), so it's dead.

Even splitting the links a little might help. Even better, you could 
break the links into "batches" and use htmerge to merge the database 
together later.

-Geoff Hutchison
Williams Students Online
http://wso.williams.edu/

------------------------------------
To unsubscribe from the htdig mailing list, send a message to
[EMAIL PROTECTED]
You'll receive a message confirming the unsubscription.

Reply via email to