Hey,

I've run into an issue which is preventing me from moving forwards with ES. 
I've got an application where I keep 'live' documents in ElasticSearch. 
Each document is a combination from data from multiple sources, which are 
merged together using doc_as_upsert. Each document has a TTL which is 
updated whenever new data comes in for a document, so documents die 
whenever no data source has given information about it for a while. The 
amount of documents generally doesn't exceed 15.000 so it's a fairly small 
data set.

Whenever I leave this running, slowly but surely memory usage on the box 
creeps up, seemingly unbounded until there is no more resident memory left. 
The Java process nicely keeps within its set ES_MAX_HEAP bounds, but it 
seems the mapping from storage on disk to memory is every-increasing, even 
when the amount of 'live' documents goes to 0. 

I was wondering if anyone has seen such a memory problem before and whether 
there are ways to debug memory usage which is unaccounted for by processes 
in 'top'.

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/68ac8858-9074-43f1-9ad4-666de8cba344%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to