Re: [ClusterLabs] Corosync on a home network

2017-09-12 Thread J Martin Rushton


On 12/09/17 19:38, Ken Gaillot wrote:
> On Mon, 2017-09-11 at 23:38 +0100, J Martin Rushton wrote:
>> I've had it switched off over the last week whilst I've been trying to
>> sort this out, but forgot tonight.  It must be the combination of
>> setting multicast_querier and stopping the firewall that is needed.  I
> 
> FYI this should open the requisite ports:
> 
> firewall-cmd --permanent --add-service=high-availability
> firewall-cmd --reload
> 
Thanks for that.  I've always used the GUI in CentOS 7, though I had
been frustrated not being able to find corosync there.  High
availability was checked on in the GUI, so I went ahead and tried things
- failure.  I then slapped my head and ran your commands on the VMS,
restarted and success.


Regards,
Martin



signature.asc
Description: OpenPGP digital signature
___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Corosync on a home network

2017-09-12 Thread Ken Gaillot
On Mon, 2017-09-11 at 23:38 +0100, J Martin Rushton wrote:
> I've had it switched off over the last week whilst I've been trying to
> sort this out, but forgot tonight.  It must be the combination of
> setting multicast_querier and stopping the firewall that is needed.  I

FYI this should open the requisite ports:

firewall-cmd --permanent --add-service=high-availability
firewall-cmd --reload

You may need to tweak that if your cluster network is not in the default
zone.

> can now see:
> 
> Quorum information
> --
> Date: Mon Sep 11 23:20:15 2017
> Quorum provider:  corosync_votequorum
> Nodes:4
> Node ID:  1
> Ring ID:  1/31156
> Quorate:  Yes
> 
> Votequorum information
> --
> Expected votes:   4
> Highest expected: 4
> Total votes:  4
> Quorum:   3
> Flags:Quorate
> 
> Membership information
> --
> Nodeid  Votes Name
>  1  1 192.168.1.2 (local)
>  2  1 192.168.1.51
>  3  1 192.168.1.52
>  4  1 192.168.1.53
> 
> which is what I wanted.
> 
> Thank you very much, I can go on to build the filesystem now.
> Martin
> 
> On 11/09/17 23:14, Leon Steffens wrote:
> > Is the firewalld service running?  Just did a quick test on my Centos 7 
> > installation and by default SSH is allowed through the firewall, but 
> > corosync cannot connect to the other nodes.
> > 
> > Try: systemctl stop firewalld.service
> 
> ___
> Users mailing list: Users@clusterlabs.org
> http://lists.clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

-- 
Ken Gaillot 





___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Corosync on a home network

2017-09-11 Thread J Martin Rushton
I've had it switched off over the last week whilst I've been trying to
sort this out, but forgot tonight.  It must be the combination of
setting multicast_querier and stopping the firewall that is needed.  I
can now see:

Quorum information
--
Date: Mon Sep 11 23:20:15 2017
Quorum provider:  corosync_votequorum
Nodes:4
Node ID:  1
Ring ID:  1/31156
Quorate:  Yes

Votequorum information
--
Expected votes:   4
Highest expected: 4
Total votes:  4
Quorum:   3
Flags:Quorate

Membership information
--
Nodeid  Votes Name
 1  1 192.168.1.2 (local)
 2  1 192.168.1.51
 3  1 192.168.1.52
 4  1 192.168.1.53

which is what I wanted.

Thank you very much, I can go on to build the filesystem now.
Martin

On 11/09/17 23:14, Leon Steffens wrote:
> Is the firewalld service running?  Just did a quick test on my Centos 7 
> installation and by default SSH is allowed through the firewall, but corosync 
> cannot connect to the other nodes.
> 
> Try: systemctl stop firewalld.service



signature.asc
Description: OpenPGP digital signature
___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] Corosync on a home network

2017-09-11 Thread Leon Steffens
Is the firewalld service running?  Just did a quick test on my Centos 7 
installation and by default SSH is allowed through the firewall, but corosync 
cannot connect to the other nodes.

Try: systemctl stop firewalld.service




> On 12 Sep 2017, at 8:04 am, J Martin Rushton  
> wrote:
> 
> Hi,
> 
> I posted the message below on the CentOS mailing list and was advised to
> repost here.  Since I posted I have also been advised to issue "echo 1 >
> /sys/class/net/br3/bridge/multicast_querier" on the main node and have
> tried it, but to no avail.
> 
> As it says in the original, any help will be gratefully received.
> 
> Regards,
> Martin
> 
> -original message--
> 
> I've been trying to build a model cluster using three virtual machines
> on my home server.  Each VM boots off its own dedicated partition
> (CentOS 7.3).  One partition is designated to be the common /home
> partition for the VMs, (on the real machine it will mount as /cluster).
> I'm intending to run GFS2 on the shared partition, so I need to
> configure DLM and corosync.  That's where I'm getting bogged down.
> 
> The VMs and the real machine are bridged onto one ethernet.  There is
> another ethernet in the main machine on a different network, but that is
> not used for clustering.  The ethernet port is connected to a switch
> which in turn connects to a BT Home Hub 6.  All four adresses are
> static, Network Manager is off, ssh works across the nodes without a
> password and ping gives sensible times.
> 
> --%<---
> # brctl show
> bridge name   bridge id   STP enabled interfaces
> br3   X   no  enp3s0
>   vnet0
>   vnet1
>   vnet2
> virbr0X   yes virbr0-nic
> --%<---
> 
> When I start corosync each node starts up but does not see the others.
> For instance I see:
> 
> --%<--
> # corosync-quorumtool
> Quorum information
> --
> Date: Sun Sep 10 12:56:56 2017
> Quorum provider:  corosync_votequorum
> Nodes:1
> Node ID:  3
> Ring ID:  3/28648
> Quorate:  No
> 
> Votequorum information
> --
> Expected votes:   4
> Highest expected: 4
> Total votes:  1
> Quorum:   3 Activity blocked
> Flags:
> 
> Membership information
> --
>Nodeid  Votes Name
> 3  1 192.168.1.52 (local)
> %<---
> 
> All four nodes are similar, but with different node IDs, IP addresses
> and Ring IDs.
> 
> The documentation warns that not all routers will handle multicast
> datagrams correctly.  I therefore attempted to force unicast
> communication by making the following changes from the distributed
> corosync.conf:
> 
>   transport: updu
>   cluster_name: 
> # crypto_cipher: none
> # crypto_hash: none
> # mcastaddr: 239.255.1.1
> # mcastport: 5405
> # ttl: 1
> 
> The following are unchanged:
> 
>   version: 2
>   secauth: off
>   ringnumber: 0
>   bindnetaddr: 192.168.1.0
> 
> The nodelist is:
> 
> -%<
> nodelist {
>   node {
>   ring0_addr: 192.168.1.2
>   nodeid: 1
>   }
>   node {
>   ring0_addr: 192.168.1.51
>   nodeid: 2
>   }
>   node {
>   ring0_addr: 192.168.1.52
>   nodeid: 3
>   }
>   node {
>   ring0_addr: 192.168.1.53
>   nodeid: 4
>   }
> }
> %<--
> 
> logging and quorum are as supplied.
> 
> Any help will be gratefully received.
> 
> Regards,
> Martin
> 
> ___
> Users mailing list: Users@clusterlabs.org
> http://lists.clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org


___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] Corosync on a home network

2017-09-11 Thread J Martin Rushton
Hi,

I posted the message below on the CentOS mailing list and was advised to
repost here.  Since I posted I have also been advised to issue "echo 1 >
/sys/class/net/br3/bridge/multicast_querier" on the main node and have
tried it, but to no avail.

As it says in the original, any help will be gratefully received.

Regards,
Martin

-original message--

I've been trying to build a model cluster using three virtual machines
on my home server.  Each VM boots off its own dedicated partition
(CentOS 7.3).  One partition is designated to be the common /home
partition for the VMs, (on the real machine it will mount as /cluster).
I'm intending to run GFS2 on the shared partition, so I need to
configure DLM and corosync.  That's where I'm getting bogged down.

The VMs and the real machine are bridged onto one ethernet.  There is
another ethernet in the main machine on a different network, but that is
not used for clustering.  The ethernet port is connected to a switch
which in turn connects to a BT Home Hub 6.  All four adresses are
static, Network Manager is off, ssh works across the nodes without a
password and ping gives sensible times.

--%<---
# brctl show
bridge name bridge id   STP enabled interfaces
br3 X   no  enp3s0
vnet0
vnet1
vnet2
virbr0  X   yes virbr0-nic
--%<---

When I start corosync each node starts up but does not see the others.
For instance I see:

--%<--
# corosync-quorumtool
Quorum information
--
Date: Sun Sep 10 12:56:56 2017
Quorum provider:  corosync_votequorum
Nodes:1
Node ID:  3
Ring ID:  3/28648
Quorate:  No

Votequorum information
--
Expected votes:   4
Highest expected: 4
Total votes:  1
Quorum:   3 Activity blocked
Flags:

Membership information
--
Nodeid  Votes Name
 3  1 192.168.1.52 (local)
%<---

All four nodes are similar, but with different node IDs, IP addresses
and Ring IDs.

The documentation warns that not all routers will handle multicast
datagrams correctly.  I therefore attempted to force unicast
communication by making the following changes from the distributed
corosync.conf:

transport: updu
cluster_name: 
#   crypto_cipher: none
#   crypto_hash: none
#   mcastaddr: 239.255.1.1
#   mcastport: 5405
#   ttl: 1

The following are unchanged:

version: 2
secauth: off
ringnumber: 0
bindnetaddr: 192.168.1.0

The nodelist is:

-%<
nodelist {
node {
ring0_addr: 192.168.1.2
nodeid: 1
}
node {
ring0_addr: 192.168.1.51
nodeid: 2
}
node {
ring0_addr: 192.168.1.52
nodeid: 3
}
node {
ring0_addr: 192.168.1.53
nodeid: 4
}
}
%<--

logging and quorum are as supplied.

Any help will be gratefully received.

Regards,
Martin



signature.asc
Description: OpenPGP digital signature
___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org