I just did an HTDump on the database that is not 'cooperating'. The dump
of the documents file seems fine.  However when I did an HTDump on the
word database I rec'd absolutely nothing, guess that would explain why
my searches aren't finging anything!  The results of the htdump -d only
gives the header line...nothing else.

My db.words.db file is 1.5G though, so I'm wondering why nothing was
dumped?  Any ideas on this one?

I didn't compile with large file support, but since the largest database
is 1.5G I shouldn't have to worry about this?  I did run htdig on a new
database I set up that had only 6000 files, and everything seems fine, I
also did an htdump on this database, and the word and document dump
files seemed to be fine.

Right now I'm working off of two years worth of data.  This year we will
be adding another, so the databases will grow by 1/3 this year.  I'd
like to keep the database searches quick, and keep the databases on the
smaller side...as the words database will definitely be larger than 2G
by the end of this year.  I'm considering splitting the database up into
two smaller databases for different filings, rather than keeping one
large database.  Any pointers/suggestions for working with large
databases such as this one would truly be appreciated!

Thanks!

Abbie


-----Original Message-----
From: Jim Cole [mailto:[EMAIL PROTECTED] 
Sent: Tuesday, March 04, 2003 8:58 PM
To: Abbie Greene
Subject: Re: [htdig] RE: No matches 3.2.0b4

On Tuesday, March 4, 2003, at 05:42 AM, Abbie Greene wrote:

> Nope, I'm receiving no error message when I do a search.  I double 
> checked to be sure that it was pointing to the write database 
> directory, and correct conf file...all seem fine.  It's strange when I

> reran rundig as root on a very small subsection of files (about 100), 
> searches came out

You might want to try the htdump utility to dump the databases to text 
files so that you can see what is there. If the words/URLs you are 
interested in show up in the dump, then in theory htsearch should be 
able to find them. If the command fails, or you end up with a bunch of 
garbage in the files, then there is a good chance that the databases 
were somehow corrupted.

> fine.  The database size is 2.3 GB total.  I did not compile for large

> file support, should this have been done and if so, how?  Also I 
> haven't found anything in regards to running htsearch from the command

> line...

Compiling for large file support requires that you specify the need for 
such support when you initially run configure. I believe the option is 
--enable-bigfile; if you do ./configure --help, you should find the 
option described in the output. This support is required for cases 
where single database files exceed the 2 GB limit. Your operating 
system (or more specifically file systems and libraries) must also 
support large files. If your largest single database file is less than 
2 GB, ht://Dig shouldn't require large file support.

> Would I just run ./htsearch [search word] like that?

If you just go to the directory and type ./htsearch and hit return, you 
will be prompted for search terms. Invoking htsearch in this way causes 
it to use the default configuration file.

Jim



-------------------------------------------------------
This SF.net email is sponsored by: Etnus, makers of TotalView, The debugger
for complex code. Debugging C/C++ programs can leave you feeling lost and
disoriented. TotalView can help you find your way. Available on major UNIX
and Linux platforms. Try it free. www.etnus.com
_______________________________________________
htdig-general mailing list <[EMAIL PROTECTED]>
To unsubscribe, send a message to <[EMAIL PROTECTED]> with a subject of unsubscribe
FAQ: http://htdig.sourceforge.net/FAQ.html

Reply via email to