I have an index with several million documents that each contains between a few hundred terms and up to about a million terms. To me it feels like there would be a rather big difference between the swetspot setting for the cache size when adding very large and very small documents.

What are the other factors I have to consider when benchmarking this? Number of threads? Initial index size?


The things is that I don't know what good the cache does in the first place nor what it does. Perhaps this is all in vain, but I'm sort of hoping it's possible to automatically find and set the cache sweetspot by sampling miscellaneous data in realtime.

Does this make sense?


        karl

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to