Currently I have 279 indexes on a single node and elasticsearch starts for 
few minutes and dies ; I only have 60G RAM on disk and as far as I know 60% 
is the max that one should allocate to elasticsearch ; I tried allocating 
38G and it lasted for few more minutes and it died. 

*(I think there's some state files that tell ES/Lucene which indexes are on 
disk)* => Where is this ? How do I fix it so that it doesn't move all 
indexes to all nodes ? I want to split the ~280 indexes into two nodes of 
140each. So far I am not able to achieve this as the master keeps moving 
nodes to itself ! 

On Monday, May 5, 2014 3:25:05 PM UTC-4, Nate Fox wrote:
>
> How many indexes do you have? It almost looks like the system itself cant 
> allocate the ram needed? 
> You might try jacking up the nofile to something like 999999 as well? I'd 
> definitely go with 31g heapsize.
>
> As for moving indexes, you might be able to copy the entire data store, 
> then remove some (I think there's some state files that tell ES/Lucene 
> which indexes are on disk), so it might recover if its missing some and 
> sees the others on another node?
>
> As for your other questions, it depends on usage as to how many nodes - 
> especially search activity while indexing. We have 230 indexes (1740 
> shards) on 8 data nodes (5.7Tb / 6.1B docs). So it can definitely handle a 
> lot more than what you're throwing at it. We dont search often nor do we 
> load a ton of data at once.
>
>
> On Sunday, May 4, 2014 7:13:09 AM UTC-7, Nish wrote:
>>
>> elasticsearch is set as a single node instance on a  60G RAM and 
>> 32*2.6GHz machine. I am actively indexing historic data with logstash. It 
>> worked well with ~300 million documents (search and indexing were doing ok) 
>> , but all of a sudden es fails to starts and keep itself up. It starts for 
>> few minutes and I can query but fails with out of memory error. I monitor 
>> the memory and atleast 12G of memory is available when it fails. I had set 
>> the es_heap_size to 31G and then reduced it to 28, 24 and 18 and the same 
>> error every time (see dump below)
>>
>> *My security limits are as under  (this is a test/POC server thus "root" 
>> user) *
>>
>> root   soft    nofile          65536
>> root   hard    nofile          65536
>> root   -       memlock         unlimited
>>
>> *ES settings *
>> config]# grep -v "^#" elasticsearch.yml | grep -v "^$"
>>  bootstrap.mlockall: true
>>
>> *echo $ES_HEAP_SIZE*
>> 18432m
>>
>> ---DUMP----
>>
>> # bin/elasticsearch
>> [2014-05-04 13:30:12,653][INFO ][node                     ] [Sabretooth] 
>> version[1.1.1], pid[19309], build[f1585f0/2014-04-16T14:27:12Z]
>> [2014-05-04 13:30:12,653][INFO ][node                     ] [Sabretooth] 
>> initializing ...
>> [2014-05-04 13:30:12,669][INFO ][plugins                  ] [Sabretooth] 
>> loaded [], sites []
>> [2014-05-04 13:30:15,390][INFO ][node                     ] [Sabretooth] 
>> initialized
>> [2014-05-04 13:30:15,390][INFO ][node                     ] [Sabretooth] 
>> starting ...
>> [2014-05-04 13:30:15,531][INFO ][transport                ] [Sabretooth] 
>> bound_address {inet[/0:0:0:0:0:0:0:0:9300]}, publish_address {inet[/
>> 10.109.136.59:9300]}
>> [2014-05-04 13:30:18,553][INFO ][cluster.service          ] [Sabretooth] 
>> new_master 
>> [Sabretooth][eocFkTYMQnSTUar94A2vHw][ip-10-109-136-59][inet[/10.109.136.59:9300]],
>>  
>> reason: zen-disco-join (elected_as_master)
>> [2014-05-04 13:30:18,579][INFO ][discovery                ] [Sabretooth] 
>> elasticsearch/eocFkTYMQnSTUar94A2vHw
>> [2014-05-04 13:30:18,790][INFO ][http                     ] [Sabretooth] 
>> bound_address {inet[/0:0:0:0:0:0:0:0:9200]}, publish_address {inet[/
>> 10.109.136.59:9200]}
>> [2014-05-04 13:30:19,976][INFO ][gateway                  ] [Sabretooth] 
>> recovered [278] indices into cluster_state
>> [2014-05-04 13:30:19,984][INFO ][node                     ] [Sabretooth] 
>> started
>> OpenJDK 64-Bit Server VM warning: Attempt to protect stack guard pages 
>> failed.
>> OpenJDK 64-Bit Server VM warning: Attempt to deallocate stack guard pages 
>> failed.
>> OpenJDK 64-Bit Server VM warning: INFO: 
>> os::commit_memory(0x00000007f7c70000, 196608, 0) failed; error='Cannot 
>> allocate memory' (errno=12)
>> #
>> # There is insufficient memory for the Java Runtime Environment to 
>> continue.
>> # Native memory allocation (malloc) failed to allocate 196608 bytes for 
>> committing reserved memory.
>> # An error report file with more information is saved as:
>> # /tmp/jvm-19309/hs_error.log
>>
>>
>>
>> ----
>> *user untergeek on #logstash told me that I have reached a max number of 
>> indices on a single node. Here are my questions: *
>>
>>    1. Can I move half of my indexes to a new node ? If yes, how to do 
>>    that without compromising indexes
>>    2. Logstash makes 1 index per day and I want to have 2 years of data 
>>    indexable ; Can I combine multiple indexes into one ? Like one month per 
>>    month : this will mean I will not have more than 24 indexes.
>>    3. How many nodes are ideal for 24 moths of data ~1.5G a day
>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/564e2951-ed54-4f34-97a9-4de88f187a7a%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to