Ok, great. Thanks Lennart. I'm not sure what I did when initially setting my elasticsearch cluster but I'm not seeing any replicas based on my viewing the bigdesk plugin and the es head plugin. I should be able to correct the number of replicas anytime, correct?
On Thursday, January 2, 2014 2:38:07 PM UTC-6, lennart wrote: > > Hey Jeremy, > > future versions of Graylog2 allow you to do this automatically but for > now you'll indeed have to move all shards over to a node you are not > going to shut down. Make sure to turn off auto-rebalancing first or > ElasticSearch will immediately start to move over other shards. > > Note that you can also configure Graylog2 to create replicas of each > shard. This way you could turn off the second node completely without > interrupting or moving anything by hand. > > Thanks, > Lennart > > On Tue, Dec 31, 2013 at 7:27 PM, Jeremy Farr > <[email protected]<javascript:>> > wrote: > > What is the recommend method for decommissioning an ES node from my > cluster? > > Obviously, I'd like to not lose any information. I've read a few posts > in > > elasticsearch forums about index shard allocation and other posts about > > simply turning the node off and allowing the cluster to rebalance. > What's > > the best or recommended method? Currently, I have used the default > settings > > in my elasticsearch.yml config as well as in the graylog2.conf. I also > only > > have two nodes in my cluster. I can easily add nodes if needed. > > > > -- > > You received this message because you are subscribed to the Google > Groups > > "graylog2" group. > > To unsubscribe from this group and stop receiving emails from it, send > an > > email to [email protected] <javascript:>. > > For more options, visit https://groups.google.com/groups/opt_out. > -- You received this message because you are subscribed to the Google Groups "graylog2" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/groups/opt_out.
