Hi Honza, I am running on the packaged version from RHEL 6.4. The exact version is 'corosync-1.4.1-15'
Thanks Lax -----Original Message----- From: linux-cluster-boun...@redhat.com [mailto:linux-cluster-boun...@redhat.com] On Behalf Of Jan Friesse Sent: Monday, November 03, 2014 1:03 AM To: linux clustering Subject: Re: [Linux-cluster] daemon cpg_join error retrying Lax, > > > >> This is just weird. What exact version of corosync are you running? Do you >> have latest Z stream? > I am running on Corosync 1.4.1 and pacemaker version is 1.1.8-7.el6 Are you running package version (like RHEL/CentOS) or did you compiled package by yourself? If package version, can you please send exact version (like 1.4.1-17.1)? > How should I get access to Z stream? Is there a specific dir I should pick > this z stream from? For RHEL you are subscribed to RHN, so you should get it automatically, for CentOS, you should get it automatically. Regards, Honza > > Thanks > Lax > > > -----Original Message----- > From: linux-cluster-boun...@redhat.com > [mailto:linux-cluster-boun...@redhat.com] On Behalf Of Jan Friesse > Sent: Friday, October 31, 2014 9:43 AM > To: linux clustering > Subject: Re: [Linux-cluster] daemon cpg_join error retrying > > Lax, > > >> Thanks Honza. Here is what I was doing, >> >>> usual reasons for this problem: >>> 1. mtu is too high and fragmented packets are not enabled (take a >>> look to netmtu configuration option) >> I am running with default mtu settings which is 1500. And I do see my >> interface(eth1) on the box does have MTU as 1500 too. >> > > Keep in mind that if they are not directly connected, switch can throw > packets because of MTU. > >> >> 2. config file on nodes are not in sync and one node may contain more node >> entries then other nodes (this may be also the case if you have two > >> clusters and one cluster contains entry of one node for other cluster) 3. >> firewall is asymmetrically blocked (so node can send but not receive). Also >> keep in mind that ports 5404 & 5405 may not be enough for udpu, because udpu >> uses one socket per remote node for sending. >> Verfiifed my config files cluster.conf and cib.xml and both have same >> no of node entries (2) >> >>> I would recommend to disable firewall completely (for testing) and if >>> everything will work, you just need to adjust firewall. >> I also ran tests with firewall off too on both the participating >> nodes, still see same issue >> >> In corosync log I see repeated set of these messages, hoping these will give >> some more pointers. >> >> Oct 29 22:11:02 corosync [SYNC ] Committing synchronization for >> (corosync cluster closed process group service v1.01) Oct 29 22:11:02 >> corosync [MAIN ] Completed service synchronization, ready to provide >> service. >> Oct 29 22:11:02 corosync [TOTEM ] waiting_trans_ack changed to 0 Oct >> 29 22:11:03 corosync [TOTEM ] entering GATHER state from 11. >> Oct 29 22:11:03 corosync [TOTEM ] entering GATHER state from 10. >> Oct 29 22:11:05 corosync [TOTEM ] entering GATHER state from 0. > > This is just weird. What exact version of corosync are you running? Do you > have latest Z stream? > > Regards, > Honza > >> Oct 29 22:11:05 corosync [TOTEM ] got commit token Oct 29 22:11:05 >> corosync [TOTEM ] Saving state aru 1b high seq received 1b Oct 29 >> 22:11:05 corosync [TOTEM ] Storing new sequence id for ring 51708 Oct >> 29 22:11:05 corosync [TOTEM ] entering COMMIT state. >> Oct 29 22:11:05 corosync [TOTEM ] got commit token Oct 29 22:11:05 >> corosync [TOTEM ] entering RECOVERY state. >> Oct 29 22:11:05 corosync [TOTEM ] TRANS [0] member 172.28.0.64: >> Oct 29 22:11:05 corosync [TOTEM ] TRANS [1] member 172.28.0.65: >> Oct 29 22:11:05 corosync [TOTEM ] position [0] member 172.28.0.64: >> Oct 29 22:11:05 corosync [TOTEM ] previous ring seq 333572 rep >> 172.28.0.64 Oct 29 22:11:05 corosync [TOTEM ] aru 1b high delivered >> 1b received flag 1 Oct 29 22:11:05 corosync [TOTEM ] position [1] member >> 172.28.0.65: >> Oct 29 22:11:05 corosync [TOTEM ] previous ring seq 333572 rep >> 172.28.0.64 Oct 29 22:11:05 corosync [TOTEM ] aru 1b high delivered >> 1b received flag 1 Oct 29 22:11:05 corosync [TOTEM ] Did not need to >> originate any messages in recovery. >> Oct 29 22:11:05 corosync [TOTEM ] token retrans flag is 0 my set >> retrans flag0 retrans queue empty 1 count 0, aru ffffffff Oct 29 >> 22:11:05 corosync [TOTEM ] install seq 0 aru 0 high seq received 0 >> Oct >> 29 22:11:05 corosync [TOTEM ] token retrans flag is 0 my set retrans >> flag0 retrans queue empty 1 count 1, aru 0 Oct 29 22:11:05 corosync >> [TOTEM ] install seq 0 aru 0 high seq received 0 Oct 29 22:11:05 >> corosync [TOTEM ] token retrans flag is 0 my set retrans flag0 >> retrans queue empty 1 count 2, aru 0 Oct 29 22:11:05 corosync [TOTEM >> ] install seq 0 aru 0 high seq received 0 Oct 29 22:11:05 corosync >> [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue >> empty 1 count 3, aru 0 Oct 29 22:11:05 corosync [TOTEM ] install seq >> 0 aru 0 high seq received 0 Oct 29 22:11:05 corosync [TOTEM ] retrans >> flag count 4 token aru 0 install seq 0 aru 0 0 Oct 29 22:11:05 >> corosync [TOTEM ] Resetting old ring state Oct 29 22:11:05 corosync >> [TOTEM ] recovery to regular 1-0 Oct 29 22:11:05 corosync [CMAN ] ais: >> confchg_fn called type = 1, seq=333576 Oct 29 22:11:05 corosync >> [TOTEM ] waiting_trans_ack changed to 1 Oct 29 22:11:05 corosync >> [CMAN ] >> ais: confchg_fn called type = 0, seq=333576 Oct 29 22:11:05 corosync >> [CMAN ] ais: last memb_count = 2, current = 2 Oct 29 22:11:05 >> corosync [CMAN ] memb: sending TRANSITION message. cluster_name = >> vsomcluster Oct 29 22:11:05 corosync [CMAN ] ais: comms send message >> 0x7fff8185ca00 len = 65 Oct 29 22:11:05 corosync [CMAN ] daemon: sending >> reply 103 to fd 24 Oct 29 22:11:05 corosync [CMAN ] daemon: sending reply >> 103 to fd 34 Oct 29 22:11:05 corosync [SYNC ] This node is within the >> primary component and will provide service. >> Oct 29 22:11:05 corosync [TOTEM ] entering OPERATIONAL state. >> Oct 29 22:11:05 corosync [TOTEM ] A processor joined or left the membership >> and a new membership was formed. >> Oct 29 22:11:05 corosync [CMAN ] ais: deliver_fn source nodeid = 2, >> len=81, endian_conv=0 Oct 29 22:11:05 corosync [CMAN ] memb: Message >> on port 0 is 5 Oct 29 22:11:05 corosync [CMAN ] memb: got TRANSITION >> from node 2 Oct 29 22:11:05 corosync [CMAN ] memb: Got TRANSITION >> message. msg->flags=20, node->flags=20, first_trans=0 Oct 29 22:11:05 >> corosync [CMAN ] memb: add_ais_node ID=2, incarnation = 333576 Oct >> 29 >> 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 corosync >> [SYNC ] Barrier Start Received From 2 Oct 29 22:11:05 corosync [SYNC ] >> Barrier completion status for nodeid 1 = 0. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [CMAN ] ais: deliver_fn source nodeid = 1, >> len=81, endian_conv=0 Oct 29 22:11:05 corosync [CMAN ] memb: Message >> on port 0 is 5 Oct 29 22:11:05 corosync [CMAN ] memb: got TRANSITION >> from node 1 Oct 29 22:11:05 corosync [CMAN ] Completed first >> transition with nodes on the same config versions Oct 29 22:11:05 >> corosync [CMAN ] memb: Got TRANSITION message. msg->flags=20, >> node->flags=20, first_trans=0 Oct 29 22:11:05 corosync [CMAN ] memb: >> add_ais_node ID=1, incarnation = 333576 Oct 29 22:11:05 corosync >> [SYNC ] confchg entries 2 Oct 29 22:11:05 corosync [SYNC ] Barrier Start >> Received From 1 Oct 29 22:11:05 corosync [SYNC ] Barrier completion status >> for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Synchronization barrier completed >> Oct 29 22:11:05 corosync [SYNC ] Synchronization actions starting >> for (dummy CLM service) Oct 29 22:11:05 corosync [SYNC ] confchg >> entries >> 2 Oct 29 22:11:05 corosync [SYNC ] Barrier Start Received From 1 Oct >> 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 0. >> Oct 29 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier Start Received From 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Synchronization barrier completed >> Oct 29 22:11:05 corosync [SYNC ] Committing synchronization for >> (dummy CLM service) Oct 29 22:11:05 corosync [SYNC ] Synchronization >> actions starting for (dummy AMF service) Oct 29 22:11:05 corosync >> [SYNC ] confchg entries 2 Oct 29 22:11:05 corosync [SYNC ] Barrier >> Start Received From 2 Oct 29 22:11:05 corosync [SYNC ] Barrier completion >> status for nodeid 1 = 0. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier Start Received From 1 Oct 29 22:11:05 >> corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Synchronization barrier completed >> Oct 29 22:11:05 corosync [SYNC ] Committing synchronization for >> (dummy AMF service) Oct 29 22:11:05 corosync [SYNC ] Synchronization >> actions starting for (openais checkpoint service B.01.01) Oct 29 >> 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 corosync >> [SYNC ] confchg entries 2 Oct 29 22:11:05 corosync [SYNC ] Barrier >> Start Received From 1 Oct 29 22:11:05 corosync [SYNC ] Barrier completion >> status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 0. >> Oct 29 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier Start Received From 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Synchronization barrier completed >> Oct 29 22:11:05 corosync [SYNC ] Committing synchronization for >> (openais checkpoint service B.01.01) Oct 29 22:11:05 corosync [SYNC >> ] Synchronization actions starting for (dummy EVT service) Oct 29 >> 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 corosync >> [SYNC ] Barrier Start Received From 2 Oct 29 22:11:05 corosync [SYNC ] >> Barrier completion status for nodeid 1 = 0. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier Start Received From 1 Oct 29 22:11:05 >> corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Synchronization barrier completed >> Oct 29 22:11:05 corosync [SYNC ] Committing synchronization for >> (dummy EVT service) Oct 29 22:11:05 corosync [SYNC ] Synchronization >> actions starting for (corosync cluster closed process group service v1.01) >> Oct 29 22:11:05 corosync [CPG ] got joinlist message from node 1 >> Oct 29 22:11:05 corosync [CPG ] comparing: sender r(0) ip(172.28.0.65) ; >> members(old:2 left:0) >> Oct 29 22:11:05 corosync [CPG ] comparing: sender r(0) ip(172.28.0.64) ; >> members(old:2 left:0) >> Oct 29 22:11:05 corosync [CPG ] chosen downlist: sender r(0) >> ip(172.28.0.64) ; members(old:2 left:0) >> Oct 29 22:11:05 corosync [CPG ] got joinlist message from node 2 >> Oct 29 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier Start Received From 1 Oct 29 22:11:05 >> corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 0. >> Oct 29 22:11:05 corosync [SYNC ] confchg entries 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier Start Received From 2 Oct 29 22:11:05 >> corosync [SYNC ] Barrier completion status for nodeid 1 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Barrier completion status for nodeid 2 = 1. >> Oct 29 22:11:05 corosync [SYNC ] Synchronization barrier completed >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[0] group:crmd\x00, >> ip:r(0) ip(172.28.0.65) , pid:9198 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[1] group:attrd\x00, >> ip:r(0) ip(172.28.0.65) , pid:9196 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[2] group:stonith-ng\x00, >> ip:r(0) ip(172.28.0.65) , pid:9194 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[3] group:cib\x00, >> ip:r(0) ip(172.28.0.65) , pid:9193 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[4] group:pcmk\x00, >> ip:r(0) ip(172.28.0.65) , pid:9187 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[5] >> group:gfs:controld\x00, ip:r(0) ip(172.28.0.65) , pid:9111 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[6] >> group:dlm:controld\x00, ip:r(0) ip(172.28.0.65) , pid:9057 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[7] >> group:fenced:default\x00, ip:r(0) ip(172.28.0.65) , pid:9040 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[8] >> group:fenced:daemon\x00, ip:r(0) ip(172.28.0.65) , pid:9040 >> Oct 29 22:11:05 corosync [CPG ] joinlist_messages[9] group:crmd\x00, >> ip:r(0) ip(172.28.0.64) , pid:14530 >> Oct 29 22:11:05 corosync [SYNC ] Committing synchronization for >> (corosync cluster closed process group service v1.01) Oct 29 22:11:05 >> corosync [MAIN ] Completed service synchronization, ready to provide >> service. >> >> Thanks >> Lax >> >> >> -----Original Message----- >> From: linux-cluster-boun...@redhat.com >> [mailto:linux-cluster-boun...@redhat.com] On Behalf Of Jan Friesse >> Sent: Thursday, October 30, 2014 1:23 AM >> To: linux clustering >> Subject: Re: [Linux-cluster] daemon cpg_join error retrying >> >>> >>>> On 30 Oct 2014, at 9:32 am, Lax Kota (lkota) <lk...@cisco.com> wrote: >>>> >>>> >>>>>> I wonder if there is a mismatch between the cluster name in cluster.conf >>>>>> and the cluster name the GFS filesystem was created with. >>>>>> How to check cluster name of GFS file system? I had similar >>>>>> configuration running fine in multiple other setups with no such issue. >>>> >>>>> I don't really recall. Hopefully someone more familiar with GFS2 can >>>>> chime in. >>>> Ok. >>>> >>>>>> >>>>>> Also one more issue I am seeing in one other setup a repeated >>>>>> flood of 'A processor joined or left the membership and a new >>>>>> membership was formed' messages for every 4secs. I am running >>>>>> with default TOTEM settings with token time out as 10 secs. Even >>>>>> after I increase the token, consensus values to be higher. It >>>>>> goes on flooding the same message after newer consensus defined time (eg: >>>>>> if I increase it to be 10secs, then I see new membership formed >>>>>> messages for every 10secs) >>>>>> >>>>>> Oct 29 14:58:10 VSM76-VSOM64 corosync[28388]: [TOTEM ] A processor >>>>>> joined or left the membership and a new membership was formed. >>>>>> Oct 29 14:58:10 VSM76-VSOM64 corosync[28388]: [CPG ] chosen >>>>>> downlist: sender r(0) ip(172.28.0.64) ; members(old:2 left:0) >>>>>> Oct 29 14:58:10 VSM76-VSOM64 corosync[28388]: [MAIN ] Completed >>>>>> service synchronization, ready to provide service. >>>>>> >>>>>> Oct 29 14:58:14 VSM76-VSOM64 corosync[28388]: [TOTEM ] A processor >>>>>> joined or left the membership and a new membership was formed. >>>>>> Oct 29 14:58:14 VSM76-VSOM64 corosync[28388]: [CPG ] chosen >>>>>> downlist: sender r(0) ip(172.28.0.64) ; members(old:2 left:0) >>>>>> Oct 29 14:58:14 VSM76-VSOM64 corosync[28388]: [MAIN ] Completed >>>>>> service synchronization, ready to provide service. >>>> >>>>> It does not sound like your network is particularly healthy. >>>>> Are you using multicast or udpu? If multicast, it might be worth >>>>> trying udpu >>>> >>>> I am using udpu and I also have firewall opened for ports 5404 & 5405. >>>> Tcpdump looks fine too, it does not complain of any issues. This is a VM >>>> envirornment and even if I switch to other node within same VM I keep >>>> getting same failure. >>> >>> Depending on what the host and VMs are doing, that might be your problem. >>> In any case, I will defer to the corosync guys at this point. >>> >> >> Lax, >> usual reasons for this problem: >> 1. mtu is too high and fragmented packets are not enabled (take a look to >> netmtu configuration option) 2. config file on nodes are not in sync and one >> node may contain more node entries then other nodes (this may be also the >> case if you have two clusters and one cluster contains entry of one node for >> other cluster) 3. firewall is asymmetrically blocked (so node can send but >> not receive). Also keep in mind that ports 5404 & 5405 may not be enough for >> udpu, because udpu uses one socket per remote node for sending. >> >> I would recommend to disable firewall completely (for testing) and if >> everything will work, you just need to adjust firewall. >> >> Regards, >> Honza >> >> >> >>>> >>>> Thanks >>>> Lax >>>> >>>> >>>> >>>> -----Original Message----- >>>> From: linux-cluster-boun...@redhat.com >>>> [mailto:linux-cluster-boun...@redhat.com] On Behalf Of Andrew >>>> Beekhof >>>> Sent: Wednesday, October 29, 2014 3:17 PM >>>> To: linux clustering >>>> Subject: Re: [Linux-cluster] daemon cpg_join error retrying >>>> >>>> >>>>> On 30 Oct 2014, at 9:06 am, Lax Kota (lkota) <lk...@cisco.com> wrote: >>>>> >>>>>> I wonder if there is a mismatch between the cluster name in cluster.conf >>>>>> and the cluster name the GFS filesystem was created with. >>>>> How to check cluster name of GFS file system? I had similar >>>>> configuration running fine in multiple other setups with no such issue. >>>> >>>> I don't really recall. Hopefully someone more familiar with GFS2 can chime >>>> in. >>>> >>>>> >>>>> Also one more issue I am seeing in one other setup a repeated >>>>> flood of 'A processor joined or left the membership and a new >>>>> membership was formed' messages for every 4secs. I am running with >>>>> default TOTEM settings with token time out as 10 secs. Even after >>>>> I increase the token, consensus values to be higher. It goes on >>>>> flooding the same message after newer consensus defined time (eg: >>>>> if I increase it to be 10secs, then I see new membership formed >>>>> messages for every >>>>> 10secs) >>>>> >>>>> Oct 29 14:58:10 VSM76-VSOM64 corosync[28388]: [TOTEM ] A processor >>>>> joined or left the membership and a new membership was formed. >>>>> Oct 29 14:58:10 VSM76-VSOM64 corosync[28388]: [CPG ] chosen downlist: >>>>> sender r(0) ip(172.28.0.64) ; members(old:2 left:0) >>>>> Oct 29 14:58:10 VSM76-VSOM64 corosync[28388]: [MAIN ] Completed >>>>> service synchronization, ready to provide service. >>>>> >>>>> Oct 29 14:58:14 VSM76-VSOM64 corosync[28388]: [TOTEM ] A processor >>>>> joined or left the membership and a new membership was formed. >>>>> Oct 29 14:58:14 VSM76-VSOM64 corosync[28388]: [CPG ] chosen downlist: >>>>> sender r(0) ip(172.28.0.64) ; members(old:2 left:0) >>>>> Oct 29 14:58:14 VSM76-VSOM64 corosync[28388]: [MAIN ] Completed >>>>> service synchronization, ready to provide service. >>>> >>>> It does not sound like your network is particularly healthy. >>>> Are you using multicast or udpu? If multicast, it might be worth >>>> trying udpu >>>> >>>>> >>>>> Thanks >>>>> Lax >>>>> >>>>> >>>>> -----Original Message----- >>>>> From: linux-cluster-boun...@redhat.com >>>>> [mailto:linux-cluster-boun...@redhat.com] On Behalf Of Andrew >>>>> Beekhof >>>>> Sent: Wednesday, October 29, 2014 2:42 PM >>>>> To: linux clustering >>>>> Subject: Re: [Linux-cluster] daemon cpg_join error retrying >>>>> >>>>> >>>>>> On 30 Oct 2014, at 8:38 am, Lax Kota (lkota) <lk...@cisco.com> wrote: >>>>>> >>>>>> Hi All, >>>>>> >>>>>> In one of my setup, I keep getting getting 'gfs_controld[10744]: daemon >>>>>> cpg_join error retrying'. I have a 2 Node setup with pacemaker and >>>>>> corosync. >>>>> >>>>> I wonder if there is a mismatch between the cluster name in cluster.conf >>>>> and the cluster name the GFS filesystem was created with. >>>>> >>>>>> >>>>>> Even after I force kill the pacemaker processes and reboot the server >>>>>> and bring the pacemaker back up, it keeps giving cpg_join error. Is >>>>>> there any way to fix this issue? >>>>>> >>>>>> >>>>>> Thanks >>>>>> Lax >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >>> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster