Re: [ClusterLabs] Corosync on a home network
On 12/09/17 19:38, Ken Gaillot wrote: > On Mon, 2017-09-11 at 23:38 +0100, J Martin Rushton wrote: >> I've had it switched off over the last week whilst I've been trying to >> sort this out, but forgot tonight. It must be the combination of >> setting multicast_querier and stopping the firewall that is needed. I > > FYI this should open the requisite ports: > > firewall-cmd --permanent --add-service=high-availability > firewall-cmd --reload > Thanks for that. I've always used the GUI in CentOS 7, though I had been frustrated not being able to find corosync there. High availability was checked on in the GUI, so I went ahead and tried things - failure. I then slapped my head and ran your commands on the VMS, restarted and success. Regards, Martin signature.asc Description: OpenPGP digital signature ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] Corosync on a home network
On Mon, 2017-09-11 at 23:38 +0100, J Martin Rushton wrote: > I've had it switched off over the last week whilst I've been trying to > sort this out, but forgot tonight. It must be the combination of > setting multicast_querier and stopping the firewall that is needed. I FYI this should open the requisite ports: firewall-cmd --permanent --add-service=high-availability firewall-cmd --reload You may need to tweak that if your cluster network is not in the default zone. > can now see: > > Quorum information > -- > Date: Mon Sep 11 23:20:15 2017 > Quorum provider: corosync_votequorum > Nodes:4 > Node ID: 1 > Ring ID: 1/31156 > Quorate: Yes > > Votequorum information > -- > Expected votes: 4 > Highest expected: 4 > Total votes: 4 > Quorum: 3 > Flags:Quorate > > Membership information > -- > Nodeid Votes Name > 1 1 192.168.1.2 (local) > 2 1 192.168.1.51 > 3 1 192.168.1.52 > 4 1 192.168.1.53 > > which is what I wanted. > > Thank you very much, I can go on to build the filesystem now. > Martin > > On 11/09/17 23:14, Leon Steffens wrote: > > Is the firewalld service running? Just did a quick test on my Centos 7 > > installation and by default SSH is allowed through the firewall, but > > corosync cannot connect to the other nodes. > > > > Try: systemctl stop firewalld.service > > ___ > Users mailing list: Users@clusterlabs.org > http://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org -- Ken Gaillot ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] Corosync on a home network
I've had it switched off over the last week whilst I've been trying to sort this out, but forgot tonight. It must be the combination of setting multicast_querier and stopping the firewall that is needed. I can now see: Quorum information -- Date: Mon Sep 11 23:20:15 2017 Quorum provider: corosync_votequorum Nodes:4 Node ID: 1 Ring ID: 1/31156 Quorate: Yes Votequorum information -- Expected votes: 4 Highest expected: 4 Total votes: 4 Quorum: 3 Flags:Quorate Membership information -- Nodeid Votes Name 1 1 192.168.1.2 (local) 2 1 192.168.1.51 3 1 192.168.1.52 4 1 192.168.1.53 which is what I wanted. Thank you very much, I can go on to build the filesystem now. Martin On 11/09/17 23:14, Leon Steffens wrote: > Is the firewalld service running? Just did a quick test on my Centos 7 > installation and by default SSH is allowed through the firewall, but corosync > cannot connect to the other nodes. > > Try: systemctl stop firewalld.service signature.asc Description: OpenPGP digital signature ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] Corosync on a home network
Is the firewalld service running? Just did a quick test on my Centos 7 installation and by default SSH is allowed through the firewall, but corosync cannot connect to the other nodes. Try: systemctl stop firewalld.service > On 12 Sep 2017, at 8:04 am, J Martin Rushton > wrote: > > Hi, > > I posted the message below on the CentOS mailing list and was advised to > repost here. Since I posted I have also been advised to issue "echo 1 > > /sys/class/net/br3/bridge/multicast_querier" on the main node and have > tried it, but to no avail. > > As it says in the original, any help will be gratefully received. > > Regards, > Martin > > -original message-- > > I've been trying to build a model cluster using three virtual machines > on my home server. Each VM boots off its own dedicated partition > (CentOS 7.3). One partition is designated to be the common /home > partition for the VMs, (on the real machine it will mount as /cluster). > I'm intending to run GFS2 on the shared partition, so I need to > configure DLM and corosync. That's where I'm getting bogged down. > > The VMs and the real machine are bridged onto one ethernet. There is > another ethernet in the main machine on a different network, but that is > not used for clustering. The ethernet port is connected to a switch > which in turn connects to a BT Home Hub 6. All four adresses are > static, Network Manager is off, ssh works across the nodes without a > password and ping gives sensible times. > > --%<--- > # brctl show > bridge name bridge id STP enabled interfaces > br3 X no enp3s0 > vnet0 > vnet1 > vnet2 > virbr0X yes virbr0-nic > --%<--- > > When I start corosync each node starts up but does not see the others. > For instance I see: > > --%<-- > # corosync-quorumtool > Quorum information > -- > Date: Sun Sep 10 12:56:56 2017 > Quorum provider: corosync_votequorum > Nodes:1 > Node ID: 3 > Ring ID: 3/28648 > Quorate: No > > Votequorum information > -- > Expected votes: 4 > Highest expected: 4 > Total votes: 1 > Quorum: 3 Activity blocked > Flags: > > Membership information > -- >Nodeid Votes Name > 3 1 192.168.1.52 (local) > %<--- > > All four nodes are similar, but with different node IDs, IP addresses > and Ring IDs. > > The documentation warns that not all routers will handle multicast > datagrams correctly. I therefore attempted to force unicast > communication by making the following changes from the distributed > corosync.conf: > > transport: updu > cluster_name: > # crypto_cipher: none > # crypto_hash: none > # mcastaddr: 239.255.1.1 > # mcastport: 5405 > # ttl: 1 > > The following are unchanged: > > version: 2 > secauth: off > ringnumber: 0 > bindnetaddr: 192.168.1.0 > > The nodelist is: > > -%< > nodelist { > node { > ring0_addr: 192.168.1.2 > nodeid: 1 > } > node { > ring0_addr: 192.168.1.51 > nodeid: 2 > } > node { > ring0_addr: 192.168.1.52 > nodeid: 3 > } > node { > ring0_addr: 192.168.1.53 > nodeid: 4 > } > } > %<-- > > logging and quorum are as supplied. > > Any help will be gratefully received. > > Regards, > Martin > > ___ > Users mailing list: Users@clusterlabs.org > http://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] Corosync on a home network
Hi, I posted the message below on the CentOS mailing list and was advised to repost here. Since I posted I have also been advised to issue "echo 1 > /sys/class/net/br3/bridge/multicast_querier" on the main node and have tried it, but to no avail. As it says in the original, any help will be gratefully received. Regards, Martin -original message-- I've been trying to build a model cluster using three virtual machines on my home server. Each VM boots off its own dedicated partition (CentOS 7.3). One partition is designated to be the common /home partition for the VMs, (on the real machine it will mount as /cluster). I'm intending to run GFS2 on the shared partition, so I need to configure DLM and corosync. That's where I'm getting bogged down. The VMs and the real machine are bridged onto one ethernet. There is another ethernet in the main machine on a different network, but that is not used for clustering. The ethernet port is connected to a switch which in turn connects to a BT Home Hub 6. All four adresses are static, Network Manager is off, ssh works across the nodes without a password and ping gives sensible times. --%<--- # brctl show bridge name bridge id STP enabled interfaces br3 X no enp3s0 vnet0 vnet1 vnet2 virbr0 X yes virbr0-nic --%<--- When I start corosync each node starts up but does not see the others. For instance I see: --%<-- # corosync-quorumtool Quorum information -- Date: Sun Sep 10 12:56:56 2017 Quorum provider: corosync_votequorum Nodes:1 Node ID: 3 Ring ID: 3/28648 Quorate: No Votequorum information -- Expected votes: 4 Highest expected: 4 Total votes: 1 Quorum: 3 Activity blocked Flags: Membership information -- Nodeid Votes Name 3 1 192.168.1.52 (local) %<--- All four nodes are similar, but with different node IDs, IP addresses and Ring IDs. The documentation warns that not all routers will handle multicast datagrams correctly. I therefore attempted to force unicast communication by making the following changes from the distributed corosync.conf: transport: updu cluster_name: # crypto_cipher: none # crypto_hash: none # mcastaddr: 239.255.1.1 # mcastport: 5405 # ttl: 1 The following are unchanged: version: 2 secauth: off ringnumber: 0 bindnetaddr: 192.168.1.0 The nodelist is: -%< nodelist { node { ring0_addr: 192.168.1.2 nodeid: 1 } node { ring0_addr: 192.168.1.51 nodeid: 2 } node { ring0_addr: 192.168.1.52 nodeid: 3 } node { ring0_addr: 192.168.1.53 nodeid: 4 } } %<-- logging and quorum are as supplied. Any help will be gratefully received. Regards, Martin signature.asc Description: OpenPGP digital signature ___ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org