Thanks for your support , I have done it, its working fine now. thank you so much .
Thanks & Regards, Shreekanta Jena On Mon, Mar 7, 2016 at 12:30 PM, Shreekant Jena <shreekant.j...@gmail.com> wrote: > Thank you for the reply... > but i am new to cluster configuration but both node were running fine > before reboot. > > Can u guide how to configure a fence device in this server. will be highly > appreciated . > > Thanks, > Shreekanta Jena > > On Sat, Mar 5, 2016 at 11:48 PM, Digimer <li...@alteeve.ca> wrote: > >> Working fencing is required. The rgmanager component waits for a >> successful fence message before beginning recovery (to prevent >> split-brains). >> >> On 05/03/16 04:47 AM, Shreekant Jena wrote: >> > secondary node >> > >> > -------------------------------------- >> > [root@Node2 ~]# cat /etc/cluster/cluster.conf >> > <?xml version="1.0"?> >> > <cluster alias="IVRS_DB" config_version="166" name="IVRS_DB"> >> > <fence_daemon clean_start="0" post_fail_delay="0" >> > post_join_delay="3"/> >> > <clusternodes> >> > <clusternode name="Node1" nodeid="1" votes="1"> >> > <fence/> >> > </clusternode> >> > <clusternode name="Node2" nodeid="2" votes="1"> >> > <fence/> >> > </clusternode> >> > </clusternodes> >> > <cman expected_votes="1" two_node="1"/> >> > <fencedevices/> >> > <rm> >> > <failoverdomains> >> > <failoverdomain name="Package1" ordered="1" >> > restricted="1"> >> > <failoverdomainnode name="Node1" >> > priority="1"/> >> > <failoverdomainnode name="Node2" >> > priority="1"/> >> > </failoverdomain> >> > </failoverdomains> >> > <resources> >> > <ip address="10.199.214.64" monitor_link="1"/> >> > </resources> >> > <service autostart="1" domain="PE51SPM1" exclusive="1" >> > name="PE51SPM1"> >> > <fs device="/dev/EI51SPM_DATA/SPIM_admin" >> > force_fsck="1" force_unmount="1" fsid="3446" fstype="ext3" >> > mountpoint="/SPIM/admin" name="admin" options="" self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/flatfile_upload" >> > force_fsck="1" force_unmount="1" fsid="17646" fstype="ext3" >> > mountpoint="/flatfile_upload" name="flatfile_upload" options="" >> > self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/oracle" >> > force_fsck="1" force_unmount="1" fsid="64480" fstype="ext3" >> > mountpoint="/oracle" name="oracle" options="" self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/SPIM_datafile_01" >> > force_fsck="1" force_unmount="1" fsid="60560" fstype="ext3" >> > mountpoint="/SPIM/datafile_01" name="datafile_01" options="" >> > self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/SPIM_datafile_02" >> > force_fsck="1" force_unmount="1" fsid="48426" fstype="ext3" >> > mountpoint="/SPIM/datafile_02" name="datafile_02" options="" >> > self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/SPIM_redolog_01" >> > force_fsck="1" force_unmount="1" fsid="54326" fstype="ext3" >> > mountpoint="/SPIM/redolog_01" name="redolog_01" options="" >> self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/SPIM_redolog_02" >> > force_fsck="1" force_unmount="1" fsid="23041" fstype="ext3" >> > mountpoint="/SPIM/redolog_02" name="redolog_02" options="" >> self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/SPIM_redolog_03" >> > force_fsck="1" force_unmount="1" fsid="46362" fstype="ext3" >> > mountpoint="/SPIM/redolog_03" name="redolog_03" options="" >> self_fence="1"/> >> > <fs device="/dev/EI51SPM_DATA/SPIM_archives_01" >> > force_fsck="1" force_unmount="1" fsid="58431" fstype="ext3" >> > mountpoint="/SPIM/archives_01" name="archives_01" options="" >> > self_fence="1"/> >> > <script file="/etc/cluster/dbstart" >> name="dbstart"/> >> > <ip ref="10.199.214.64"/> >> > </service> >> > </rm> >> > </cluster> >> > >> > >> > [root@Node2 ~]# clustat >> > msg_open: Invalid argument >> > Member Status: Inquorate >> > >> > Resource Group Manager not running; no service information available. >> > >> > Membership information not available >> > >> > >> > >> > Primary Node >> > >> > ----------------------------------------- >> > [root@Node1 ~]# clustat >> > Member Status: Quorate >> > >> > Member Name Status >> > ------ ---- ------ >> > Node1 Online, Local, rgmanager >> > Node2 Offline >> > >> > Service Name Owner (Last) State >> > ------- ---- ----- ------ ----- >> > Package1 Node1 started >> > >> > >> > On Sat, Mar 5, 2016 at 12:17 PM, Digimer <li...@alteeve.ca >> > <mailto:li...@alteeve.ca>> wrote: >> > >> > Please share your cluster.conf (only obfuscate passwords please) >> and the >> > output of 'clustat' from each node. >> > >> > digimer >> > >> > On 05/03/16 01:46 AM, Shreekant Jena wrote: >> > > Dear All, >> > > >> > > I have a 2 node cluster but after reboot secondary node is showing >> > > offline . And cman failed to start . >> > > >> > > Please find below logs on secondary node:- >> > > >> > > root@EI51SPM1 cluster]# clustat >> > > msg_open: Invalid argument >> > > Member Status: Inquorate >> > > >> > > Resource Group Manager not running; no service information >> available. >> > > >> > > Membership information not available >> > > [root@EI51SPM1 cluster]# tail -10 /var/log/messages >> > > Feb 24 13:36:23 EI51SPM1 ccsd[25487]: Error while processing >> connect: >> > > Connection refused >> > > Feb 24 13:36:23 EI51SPM1 kernel: CMAN: sending membership request >> > > Feb 24 13:36:27 EI51SPM1 ccsd[25487]: Cluster is not quorate. >> > Refusing >> > > connection. >> > > Feb 24 13:36:27 EI51SPM1 ccsd[25487]: Error while processing >> connect: >> > > Connection refused >> > > Feb 24 13:36:28 EI51SPM1 kernel: CMAN: sending membership request >> > > Feb 24 13:36:32 EI51SPM1 ccsd[25487]: Cluster is not quorate. >> > Refusing >> > > connection. >> > > Feb 24 13:36:32 EI51SPM1 ccsd[25487]: Error while processing >> connect: >> > > Connection refused >> > > Feb 24 13:36:32 EI51SPM1 ccsd[25487]: Cluster is not quorate. >> > Refusing >> > > connection. >> > > Feb 24 13:36:32 EI51SPM1 ccsd[25487]: Error while processing >> connect: >> > > Connection refused >> > > Feb 24 13:36:33 EI51SPM1 kernel: CMAN: sending membership request >> > > [root@EI51SPM1 cluster]# >> > > [root@EI51SPM1 cluster]# cman_tool status >> > > Protocol version: 5.0.1 >> > > Config version: 166 >> > > Cluster name: IVRS_DB >> > > Cluster ID: 9982 >> > > Cluster Member: No >> > > Membership state: Joining >> > > [root@EI51SPM1 cluster]# cman_tool nodes >> > > Node Votes Exp Sts Name >> > > [root@EI51SPM1 cluster]# >> > > [root@EI51SPM1 cluster]# >> > > >> > > >> > > Thanks & regards >> > > SHREEKANTA JENA >> > > >> > > >> > > >> > >> > >> > -- >> > Digimer >> > Papers and Projects: https://alteeve.ca/w/ >> > What if the cure for cancer is trapped in the mind of a person >> without >> > access to education? >> > >> > -- >> > Linux-cluster mailing list >> > Linux-cluster@redhat.com <mailto:Linux-cluster@redhat.com> >> > https://www.redhat.com/mailman/listinfo/linux-cluster >> > >> > >> > >> > >> >> >> -- >> Digimer >> Papers and Projects: https://alteeve.ca/w/ >> What if the cure for cancer is trapped in the mind of a person without >> access to education? >> >> -- >> Linux-cluster mailing list >> Linux-cluster@redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > >
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster