Are you closing the searcher after each when done? No: Waiting for the garbage collector is not a good idea.
Yes: It could be a timeout on the OS holding the files handles. Either way, the only real option is to avoid thrashing the searchers... Scott > -----Original Message----- > From: Hang Li [mailto:[EMAIL PROTECTED]] > Sent: Tuesday, July 23, 2002 10:10 AM > To: Lucene Users List > Subject: Re: Too many open files? > > > Thanks for your quick reponse, I still want to know why we ran out of > file descriptors. > > --Yup. Cache and reuse your Searcher as much as possible. > > --Scott > > > -----Original Message----- > > From: Hang Li [mailto:[EMAIL PROTECTED]] > > Sent: Tuesday, July 23, 2002 9:59 AM > > To: Lucene Users List > > Subject: Too many open files? > > > > > > > > > > > I have seen a lot postings about this topic. Any final thoughts? > > > > We did a simple stress test, Lucene would produce this error > > between 30 - 80 > > concurren searches. The index directory has 24 files (15 > fields), and > > > " > > ulimit -n > > 32768 > > ", > > > > there should be more than enough FDs. Note, we did not do > > any writings to index > > while we were searching. Any ideas? Thx. > > > > > > -- > To unsubscribe, e-mail: <mailto:[EMAIL PROTECTED]> For additional commands, e-mail: <mailto:[EMAIL PROTECTED]>
