Hi,

I have experienced same behavior when I have tried to load large amount of 
data... If you clear the file system cache 
(here<http://www.delphitools.info/2013/11/29/flush-windows-file-cache/>is a 
link to a tool), the memory drops to the defined heap size.
However this is still looks as a wrong behavior, is there a way to block 
the shareable memory upfront ?

All the best,
Yitzhak

On Tuesday, March 18, 2014 12:34:33 AM UTC+2, Jos Kraaijeveld wrote:

> As a follow-up, when the server is nearing maximum memory, the memory use 
> stops increasing. This would indeed support Zachary's caching theory, 
> although I'm still confused as to why it shows up as 'in use' memory rather 
> than 'cached' memory. In any case, it does not block me right now. It's 
> just peculiar, and I'll revive this thread once I have a better explanation.
>
> On Thursday, March 13, 2014 5:35:17 PM UTC-7, Jos Kraaijeveld wrote:
>>
>> There are no other processes running except for ES and the program which 
>> posts the updates. The memory is constantly increasing when the updater is 
>> running, but is stale (and doesn't release the memory at all, no matter how 
>> much is used) whenever ES is idle.
>>
>> On Thursday, March 13, 2014 5:32:43 PM UTC-7, Zachary Tong wrote:
>>>
>>> Also, are there other processes running which may be causing the 
>>> problem?  Does the behavior only happen when ES is running?
>>>
>>>
>>>
>>> On Thursday, March 13, 2014 8:31:18 PM UTC-4, Zachary Tong wrote:
>>>>
>>>> Cool, curious to see what happens.  As an aside, I would recommend 
>>>> downgrading to Java 1.7.0_u25.  There are known bugs in the most recent 
>>>> Oracle JVM versions which have not been resolved yet.  u25 is the most 
>>>> recent safe version.  I don't think that's your problem, but it's a good 
>>>> general consideration anyway.
>>>>
>>>> -Z
>>>>
>>>>
>>>>
>>>> On Thursday, March 13, 2014 8:23:34 PM UTC-4, Jos Kraaijeveld wrote:
>>>>>
>>>>> @Mark:
>>>>> The heap is set to 2GB, using mlockall. The problem occurs with both 
>>>>> OpenJDK7 and OracleJDK7, both the latest versions. I have one index, 
>>>>> which 
>>>>> is very small:
>>>>> index: 
>>>>> {
>>>>> primary_size_in_bytes: 37710681
>>>>> size_in_bytes: 37710681
>>>>> }
>>>>>
>>>>> @Zachary Our systems are set up to alert when memory is about to run 
>>>>> out. We use Ganglia to monitor our systems and that represents the memory 
>>>>> as 'used', rather than 'cached'. I will try to just let it run until 
>>>>> memory 
>>>>> runs out and report back after that though.
>>>>>
>>>>>
>>>>>
>>>>> On Thursday, March 13, 2014 5:17:20 PM UTC-7, Zachary Tong wrote:
>>>>>>
>>>>>> I believe you are just witnessing the OS caching files in memory. 
>>>>>>  Lucene (and therefore by extension Elasticsearch) uses a large number 
>>>>>> of 
>>>>>> files to represent segments.  TTL + updates will cause even higher file 
>>>>>> turnover than usual.
>>>>>>
>>>>>> The OS manages all of this caching and will reclaim it for other 
>>>>>> processes when needed.  Are you experiencing problems, or just 
>>>>>> witnessing 
>>>>>> memory usage?  I wouldn't be concerned unless there is an actual problem 
>>>>>> that you are seeing.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thursday, March 13, 2014 8:07:13 PM UTC-4, Jos Kraaijeveld wrote:
>>>>>>>
>>>>>>> Hey,
>>>>>>>
>>>>>>> I've run into an issue which is preventing me from moving forwards 
>>>>>>> with ES. I've got an application where I keep 'live' documents in 
>>>>>>> ElasticSearch. Each document is a combination from data from multiple 
>>>>>>> sources, which are merged together using doc_as_upsert. Each document 
>>>>>>> has a 
>>>>>>> TTL which is updated whenever new data comes in for a document, so 
>>>>>>> documents die whenever no data source has given information about it 
>>>>>>> for a 
>>>>>>> while. The amount of documents generally doesn't exceed 15.000 so it's 
>>>>>>> a 
>>>>>>> fairly small data set.
>>>>>>>
>>>>>>> Whenever I leave this running, slowly but surely memory usage on the 
>>>>>>> box creeps up, seemingly unbounded until there is no more resident 
>>>>>>> memory 
>>>>>>> left. The Java process nicely keeps within its set ES_MAX_HEAP bounds, 
>>>>>>> but 
>>>>>>> it seems the mapping from storage on disk to memory is 
>>>>>>> every-increasing, 
>>>>>>> even when the amount of 'live' documents goes to 0. 
>>>>>>>
>>>>>>> I was wondering if anyone has seen such a memory problem before and 
>>>>>>> whether there are ways to debug memory usage which is unaccounted for 
>>>>>>> by 
>>>>>>> processes in 'top'.
>>>>>>>
>>>>>>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/82a08e2d-380f-4a03-bff5-0255a93c637d%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to