Hi,

Thanks for the configuration docu.

Can I really run into split brain?
I have 3 nodes, they are all equal. Everyone of them can be a master and 
will store data. With the discovery.zen.minimum_master_nodes: 2 I can't get 
a split brain. Or am I wrong?
Or is this setup not ideal?

Chris...

On Tuesday, February 10, 2015 at 1:38:06 PM UTC+1, Arie wrote:
>
> You coud bump into a split brain situation running all ES nodes as master.
>
> Check out this to configure your cluster:
>
>
> http://www.elasticsearch.org/guide/en/elasticsearch/guide/current/_important_configuration_changes.html#_minimum_master_nodes
>
>
>
> On Tuesday, February 10, 2015 at 12:09:33 AM UTC+1, Christoph Fürstaller 
> wrote:
>>
>> Thanks for your answer!
>>
>> About the master/data nodes. What happens when the master goes down? Will 
>> one of the 'slaves' become a master? I configured all 3 as master for 
>> redundancy, so the cluster still survives if only one node is present. Is 
>> this assumption wrong?
>>
>> I've increased the ES_HEAP_SIZE to 6G before, with the same results. 
>>
>> Chris...
>>
>> Am Montag, 9. Februar 2015 20:30:28 UTC+1 schrieb Arie:
>>>
>>> Hi,,
>>>
>>> Looking @ your config in elasticsearch.yml the follwing comes in to mind
>>>
>>> One node should be:
>>> node.master: true
>>> node.data: true
>>>  
>>> and for the other two nodes:
>>> node.master: false
>>> node.data: false
>>>
>>> elasticseaarch.conf
>>> ES_HEAP_SIZE
>>>
>>> you can take this easy up o 8G (50% of your memory) and check if this is 
>>> really
>>> running so. In my case on Centos6 I put this in 
>>> /etc/conf.d/elasticseaarch
>>>
>>> Good luck.
>>>
>>> On Friday, February 6, 2015 at 12:58:27 PM UTC+1, Christoph Fürstaller 
>>> wrote:
>>>>
>>>> Hi,
>>>> Yesterday we've updatet our Graylog2/Elasticsearch Cluster. The Elastic 
>>>> Search Cluster consists of 3 physical Maschins: DL380 G7, E5620, 16GB RAM 
>>>> on RHEL 6.6. Each ES Node gets 4GB RAM. On one Host there is the graylog2 
>>>> Server/Interface installed. Until yesterday we used Elasticsearch 
>>>> 0.90.10-1 
>>>> and graylog2-0.20.3 Yesterday we updatet graylog2 to 0.90.0, startet 
>>>> everything, everything was running fine. Then Stopped graylog2 and the 
>>>> ElaticSearch Cluster, upgraded ES to 1.3.4 and graylog to 0.92.4. The 
>>>> Upgrade from ES was successfully, after that, startet graylog2, which 
>>>> connected to the cluster and showed everything.
>>>>
>>>> In the ES Cluster there are 7 indices a 20mio messages. The last 3 
>>>> indices are opened, the other closed. Graylog2 sees approx 50mio messages. 
>>>> New messages arrive with approx 5msg/sec
>>>>
>>>> In the logs from graylog2-server there are messages like this, every 
>>>> couple of minutes:
>>>> org.graylog2.periodical.GarbageCollectionWarningThread - Last GC run 
>>>> with PS Scavenge took longer than 1 second
>>>>
>>>> It seems graylog is running fine, a bit slow on searches, but fine.
>>>>
>>>> Attached are the config files for graylog2 and elasticsearch.
>>>>
>>>> Can someone give us a hint where this warnings come from? What we can 
>>>> tweak? Would be very helpful!
>>>>
>>>> Thanks!
>>>> Chris...
>>>>
>>>

-- 
You received this message because you are subscribed to the Google Groups 
"graylog2" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to