Ryan are you able to update to that commit I just did?  If so I think
you should run the tool without -fix and post back what it printed.  It
should report an error on that one segment due to the missing file.
Then, run -fix to remove that segment (please backup your index first!).
Then, if you have a zillion segments in the index, try optimizing it?


uggg. Thanks for all your help / pointers. I've been able to salvage a functioning index - (with 800K fewer docs) - I'll just go back to an old index and build from there.

Optimizing reduced 180K files to 1800 - I'm guessing that had something to do with why it hit a max open file limit.

I'm running a standard solr-trunk from a few weeks back (nothing custom, no 'embedded' stuff). The index is constantly updating a few files, so I am not surprised by the mass of unused files optimize cleans up.

If I understand Yonik's comments, something in lucene 2.3 has changed so that hitting the max open file limit could not result in a corrupt index. Is that true? In the meantime would optimizing more often help avoid this?

thanks
ryan






---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to