Teerapatr > Dear Honza, > > Sorry to say this, but I found new error again. LOL > > This time, I already install the 1.4.1-17 as your advice. > And the nodename, without altname, is map to IPv6 using hosts file. > Everything is fine, but the 2 node can't communicate to each other. > So I add the multicast address manually, using command `ccs -f > /etc/cluster/cluster.conf --setmulticast ff::597` on both node. > After that the CMAN cannot start.
ff:: is not valid ipv6 multicast address. Use something like ff3e::597. > > Starting cluster: > Checking if cluster has been disabled at boot... [ OK ] > Checking Network Manager... [ OK ] > Global setup... [ OK ] > Loading kernel modules... [ OK ] > Mounting configfs... [ OK ] > Starting cman... Timed-out waiting for cluster Check cluster logs for > details > [FAILED] > > I also found a lot of LOG, but I think that this is where the problem has > occur. > > Jul 15 13:36:14 corosync [MAIN ] Corosync Cluster Engine ('1.4.1'): > started and ready to provide service. > Jul 15 13:36:14 corosync [MAIN ] Corosync built-in features: nss dbus rdma > snmp > Jul 15 13:36:14 corosync [MAIN ] Successfully read config from > /etc/cluster/cluster.conf > Jul 15 13:36:14 corosync [MAIN ] Successfully parsed cman config > Jul 15 13:36:14 corosync [TOTEM ] Initializing transport (UDP/IP Multicast). > Jul 15 13:36:14 corosync [TOTEM ] Initializing transmit/receive > security: libtomcrypt SOBER128/SHA1HMAC (mode 0). > Jul 15 13:36:14 corosync [TOTEM ] Unable to bind the socket to receive > multicast packets: Cannot assign requested address (99) > Jul 15 13:36:14 corosync [TOTEM ] Could not set traffic priority: > Socket operation on non-socket (88) > Jul 15 13:36:14 corosync [TOTEM ] The network interface > [2001:db8::151] is now up. > Jul 15 13:36:14 corosync [QUORUM] Using quorum provider quorum_cman > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > cluster quorum service v0.1 > Jul 15 13:36:14 corosync [CMAN ] CMAN 3.0.12.1 (built Apr 14 2014 > 09:36:10) started > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync CMAN > membership service 2.90 > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: openais > checkpoint service B.01.01 > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > extended virtual synchrony service > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > configuration service > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > cluster closed process group service v1.01 > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > cluster config database access v1.01 > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > profile loading service > Jul 15 13:36:14 corosync [QUORUM] Using quorum provider quorum_cman > Jul 15 13:36:14 corosync [SERV ] Service engine loaded: corosync > cluster quorum service v0.1 > Jul 15 13:36:14 corosync [MAIN ] Compatibility mode set to whitetank. > Using V1 and V2 of the synchronization engine. > Jul 15 13:36:17 corosync [MAIN ] Totem is unable to form a cluster > because of an operating system or network fault. The most common cause > of this message is that the local firewall is configured improperly. > Jul 15 13:36:19 corosync [MAIN ] Totem is unable to form a cluster > because of an operating system or network fault. The most common cause > of this message is that the local firewall is configured improperly. > Jul 15 13:36:20 corosync [MAIN ] Totem is unable to form a cluster > because of an operating system or network fault. The most common cause > of this message is that the local firewall is configured improperly. > > I cannot find the solution on Internet about "[TOTEM ] Unable to bind > the socket to receive multicast packets: Cannot assign requested > address (99)". > Do you have any idea? > > Teenigma > > On Tue, Jul 15, 2014 at 10:02 AM, Teerapatr Kittiratanachai > <maillist...@gmail.com> wrote: >> Honza >> >> Great, Thank you very much. >> >> But the terrible thing for me is I'm using the package from OpenSUSE repo. >> When i turn back to CentOS repo, which store lower version, the >> Dependency problem has occurred. >> >> Anyway, thank you for your help. >> >> Teenigma >> >> On Mon, Jul 14, 2014 at 8:51 PM, Jan Friesse <jfrie...@redhat.com> wrote: >>>> Honza, >>>> >>>> How do I include the patch with my CentOS package? >>>> Do I need to compile them manually? >>> >>> >>> Yes. Also official CentOS version was never 1.4.5. If you are using CentOS, >>> just use stock 1.4.1-17.1. Patch is included there. >>> >>> Honza >>> >>> >>>> >>>> TeEniGMa >>>> >>>> On Mon, Jul 14, 2014 at 3:21 PM, Jan Friesse <jfrie...@redhat.com> wrote: >>>>> >>>>> Teerapatr, >>>>> >>>>> >>>>>> For more information, >>>>>> >>>>>> >>>>>> these are LOG from /var/log/messages >>>>>> ... >>>>>> Jul 14 10:28:07 wh00 kernel: : DLM (built Mar 25 2014 20:01:13) >>>>>> installed >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [MAIN ] Corosync Cluster >>>>>> Engine ('1.4.5'): started and ready to provide service. >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [MAIN ] Corosync built-in >>>>>> features: nss >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [MAIN ] Successfully read >>>>>> config from /etc/cluster/cluster.conf >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [MAIN ] Successfully parsed cman >>>>>> config >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [TOTEM ] Initializing transport >>>>>> (UDP/IP Multicast). >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [TOTEM ] Initializing >>>>>> transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0). >>>>>> Jul 14 10:28:07 wh00 corosync[2716]: [TOTEM ] The network interface is >>>>>> down. >>>>> >>>>> >>>>> ^^^ This line is important. This means, corosync was unable to find >>>>> interface with given IPv6 address. There was regression in v1.4.5 causing >>>>> this behavior. It's fixed in v1.4.6 (patch is >>>>> >>>>> https://github.com/corosync/corosync/commit/d76759ec26ecaeb9cc01f49e9eb0749b61454d27). >>>>> So you can ether apply patch or (recommended) upgrade to 1.4.7. >>>>> >>>>> Regards, >>>>> Honza >>>>> >>>>> >>>>> >>>>>> Jul 14 10:28:10 wh00 pacemaker: Aborting startup of Pacemaker Cluster >>>>>> Manager >>>>>> ... >>>>>> >>>>>> Te >>>>>> >>>>>> On Mon, Jul 14, 2014 at 10:07 AM, Teerapatr Kittiratanachai >>>>>> <maillist...@gmail.com> wrote: >>>>>>> >>>>>>> >>>>>>> Dear Honza, >>>>>>> >>>>>>> Sorry for late reply. >>>>>>> After I have tested with all new configuration. >>>>>>> On IPv6 only, and with no altname. >>>>>>> >>>>>>> I face with error below, >>>>>>> >>>>>>> Starting cluster: >>>>>>> Checking if cluster has been disabled at boot... [ OK ] >>>>>>> Checking Network Manager... [ OK ] >>>>>>> Global setup... [ OK ] >>>>>>> Loading kernel modules... [ OK ] >>>>>>> Mounting configfs... [ OK ] >>>>>>> Starting cman... corosync died with signal: 6 Check cluster logs >>>>>>> for >>>>>>> details >>>>>>> [FAILED] >>>>>>> >>>>>>> And, exactly, there are no any enabled firewall, I also configure the >>>>>>> Multicast address as manual. >>>>>>> Could you advise me the solution? >>>>>>> >>>>>>> Many thanks in advance. >>>>>>> Te >>>>>>> >>>>>>> On Thu, Jul 10, 2014 at 6:14 PM, Jan Friesse <jfrie...@redhat.com> >>>>>>> wrote: >>>>>>>> >>>>>>>> >>>>>>>> Teerapatr, >>>>>>>> >>>>>>>>> Hi Honza, >>>>>>>>> >>>>>>>>> As you said I use the nodename identify by hostname (which be >>>>>>>>> accessed >>>>>>>>> via IPv6) and the node also has the altname (which be IPv4 address). >>>>>>>>> >>>>>>>> >>>>>>>> This doesn't work. Both hostname and altname have to be same IP >>>>>>>> version. >>>>>>>> >>>>>>>>> Now, I configure the mcast address for both nodename and altname >>>>>>>>> manually. The CMAN and Pacemaker can start ad well. But they don't >>>>>>>>> communicate to another node. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> PLease make sure (as I've wrote in previous email) your firewall >>>>>>>> doesn't >>>>>>>> block mcast and corosync traffic (just disable it) and switch doesn't >>>>>>>> block multicast (this is very often the case). If these are VMs, make >>>>>>>> sure to properly configure bridge (just disable firewall) and allow >>>>>>>> mcast_querier. >>>>>>>> >>>>>>>> Honza >>>>>>>> >>>>>>>>> On node0, crm_mon show node1 offline. In the same way, node one show >>>>>>>>> node0 is down. So the split brain problem occur here. >>>>>>>>> >>>>>>>>> Regards, >>>>>>>>> Te >>>>>>>>> >>>>>>>>> On Thu, Jul 10, 2014 at 2:50 PM, Jan Friesse <jfrie...@redhat.com> >>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Teerapatr, >>>>>>>>>> >>>>>>>>>>> OK, some problems are solved. >>>>>>>>>>> I use the incorrect hostname. >>>>>>>>>>> >>>>>>>>>>> For now, the new problem has occured. >>>>>>>>>>> >>>>>>>>>>> Starting cman... Node address family does not match multicast >>>>>>>>>>> address family >>>>>>>>>>> Unable to get the configuration >>>>>>>>>>> Node address family does not match multicast address family >>>>>>>>>>> cman_tool: corosync daemon didn't start Check cluster logs for >>>>>>>>>>> details >>>>>>>>>>> >>>>>>>>>>> [FAILED] >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> This looks like one of your node is also reachable via ipv4 and ipv4 >>>>>>>>>> resolving is proffered. Please make sure to set only ipv6 address >>>>>>>>>> and >>>>>>>>>> try it again. Of course set mcast addr by hand maybe helpful >>>>>>>>>> (even-tho >>>>>>>>>> I >>>>>>>>>> don't believe it will solve problem you are hitting)). >>>>>>>>>> >>>>>>>>>> Also make sure ip6tables are properly configured and your switch is >>>>>>>>>> able >>>>>>>>>> to pass ipv6 mcast traffic. >>>>>>>>>> >>>>>>>>>> Regards, >>>>>>>>>> Honza >>>>>>>>>> >>>>>>>>>>> How can i fix it? Or just assigned the multicast address in the >>>>>>>>>>> configuration? >>>>>>>>>>> >>>>>>>>>>> Regards, >>>>>>>>>>> Te >>>>>>>>>>> >>>>>>>>>>> On Thu, Jul 10, 2014 at 7:52 AM, Teerapatr Kittiratanachai >>>>>>>>>>> <maillist...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> I not found any LOG message >>>>>>>>>>>> >>>>>>>>>>>> /var/log/messages >>>>>>>>>>>> ... >>>>>>>>>>>> Jul 10 07:44:19 nwh00 kernel: : DLM (built Jun 19 2014 21:16:01) >>>>>>>>>>>> installed >>>>>>>>>>>> Jul 10 07:44:22 nwh00 pacemaker: Aborting startup of Pacemaker >>>>>>>>>>>> Cluster Manager >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> and this is what display when I try to start pacemaker >>>>>>>>>>>> >>>>>>>>>>>> # /etc/init.d/pacemaker start >>>>>>>>>>>> Starting cluster: >>>>>>>>>>>> Checking if cluster has been disabled at boot... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Checking Network Manager... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Global setup... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Loading kernel modules... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Mounting configfs... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Starting cman... Cannot find node name in cluster.conf >>>>>>>>>>>> Unable to get the configuration >>>>>>>>>>>> Cannot find node name in cluster.conf >>>>>>>>>>>> cman_tool: corosync daemon didn't start Check cluster logs for >>>>>>>>>>>> details >>>>>>>>>>>> >>>>>>>>>>>> [FAILED] >>>>>>>>>>>> Stopping cluster: >>>>>>>>>>>> Leaving fence domain... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Stopping gfs_controld... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Stopping dlm_controld... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Stopping fenced... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Stopping cman... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Unloading kernel modules... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Unmounting configfs... [ OK >>>>>>>>>>>> ] >>>>>>>>>>>> Aborting startup of Pacemaker Cluster Manager >>>>>>>>>>>> >>>>>>>>>>>> another one thing, according to the happened problem, I remove the >>>>>>>>>>>> AAAA record from DNS for now and map it in to /etc/hosts files >>>>>>>>>>>> instead, as shown below. >>>>>>>>>>>> >>>>>>>>>>>> /etc/hosts >>>>>>>>>>>> ... >>>>>>>>>>>> 2001:db8:0:1::1 node0.example.com >>>>>>>>>>>> 2001:db8:0:1::2 node1.example.com >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> Is there any configure that help me to got more log ? >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Jul 10, 2014 at 5:06 AM, Andrew Beekhof >>>>>>>>>>>> <and...@beekhof.net> >>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On 9 Jul 2014, at 9:15 pm, Teerapatr Kittiratanachai >>>>>>>>>>>>> <maillist...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Dear All, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I has implemented the HA on dual stack servers, >>>>>>>>>>>>>> Firstly, I doesn't deploy IPv6 record on DNS yet. The CMAN and >>>>>>>>>>>>>> PACEMAKER can work as normal. >>>>>>>>>>>>>> But, after I create AAAA record on DNS server, i found the error >>>>>>>>>>>>>> that >>>>>>>>>>>>>> cann't start CMAN. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Are CMAN and PACEMAKER support the IPv6? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> I don;t think pacemaker cares. >>>>>>>>>>>>> What errors did you get? >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>>>> >>>>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>>>> Getting started: >>>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>> >>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>> Getting started: >>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>> >>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>> Getting started: >>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>> >>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>> Getting started: >>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>> >>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>> Getting started: >>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>> >>>>>> Project Home: http://www.clusterlabs.org >>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>> Bugs: http://bugs.clusterlabs.org >>>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>> >>>>> Project Home: http://www.clusterlabs.org >>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> Bugs: http://bugs.clusterlabs.org >>>> >>>> >>>> _______________________________________________ >>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org >>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>>> >>> >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org