Dear  Emmanuel Segura,

Find the config below.  Because of policy I have removed some login details.

#############

<?xml version="1.0"?>
<cluster config_version="6" name="new_cluster">
<fence_daemon post_fail_delay="0" post_join_delay="3"/>
<clusternodes>
<clusternode name="mailbox1" nodeid="1" votes="1">
<multicast addr="224.0.0.1" interface="bond0"/>
<fence>
<method name="1">
<device name="imap1drac"/>
</method>
</fence>
</clusternode>
<clusternode name="mailbox2" nodeid="2" votes="1">
<multicast addr="224.0.0.1" interface="bond0"/>
<fence>
<method name="1">
<device name="imap2drac"/>
</method>
</fence>
</clusternode>
</clusternodes>
<cman expected_votes="1" two_node="1">
<multicast addr="224.0.0.1"/>
</cman>
<fencedevices>
<fencedevice agent="fence_drac6" ipaddr="<drac IP>" login="<login name>" name="imap1drac" passwd="xxxxx"/> <fencedevice agent="fence_drac6" ipaddr="<drac IP>" login="<login name>" name="imap2drac" passwd="xxxxx"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources>
<ip address="192.168.1.1" monitor_link="1"/>
<fs device="/dev/drbd0" force_fsck="0" force_unmount="1" fsid="28418" fstype="ext3" mountpoint="/mount/path" name="imap1_fs" options="rw" self_fence="1"/>
<script file="/etc/init.d/cyrus-imapd" name="imap1_init"/>
</resources>
<service autostart="1" name="imap1" recovery="restart">
<ip ref="192.168.1.1"/>
<fs ref="imap1_fs"/>
<script ref="imap1_init"/>
</service>
</rm>
</cluster>
###################

Regards
Jayesh Shinde


On 01/25/2012 01:59 PM, emmanuel segura wrote:
Can you show me your cluster config?

2012/1/25 jayesh.shinde <jayesh.shi...@netcore.co.in <mailto:jayesh.shi...@netcore.co.in>>

    Hi  all ,

    I have few queries about fence working.

    I am using 2 different  the 2 node cluster with Dell and IBM
    hardware in two different IDC.
    Recently I came across the network failure problem at different
    time and I found my 2 nodes are power off state.

    Below is  how the situation happened with my 2 different 2 node
    cluster.

    With 2 node IBM node cluster with SAN :--
    ==============================
    1)  Network connectivity  was failed totally for few minutes.
    2) And as per the /var/log/messages both servers failed to  fence
    to each other and both server was UP as it is with all services.
    3) But the "clustat" was showing serves are not in cluster mode
    and "regmanger" status was stop.
    4) I simply reboot the server.
    5) After that I found both server in power off stat.


    with another  2 node Dell server with DRBD  :--
    =================================
    1) Network connectivity  was failed totally.
    2) DRAC ip was unavailable so fence failed from both server.
    3) after some time I fond the servers are shutdown.

    In normal conditions both cluster work properly

     my queries are now :--
     ===============
    1) What could be the reason for power off ?
    2) Does cluster's fencing method  caused for the power off  of
    server ( i.e because of previous failed fence ) ?
    3) Is there any test cases mentioned on net / blog / wiki  about
    the fence , i.e different situation under which  fence works.

    Please guide.

    Thanks & Regards
    Jayesh Shinde




    --
    Linux-cluster mailing list
    Linux-cluster@redhat.com <mailto:Linux-cluster@redhat.com>
    https://www.redhat.com/mailman/listinfo/linux-cluster




--
esta es mi vida e me la vivo hasta que dios quiera


--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to