Carlopmart,
it's not problem of fence_vmware_ng and definitively not malfunction.
Cluster is trying to keep all possible nodes going. So what will happend:
- You will boot node01
- This will start fencing, and finds out, that node02 isn't running (is
falled, ...) -> fence node2
- Node2 is fenced and starts

Regards,
  Honza

carlopmart wrote:
> Hi all,
> 
>  I have installed two rhel5.4 nodes virtual guests, el5prodnode01 and
> el5prodnode02, under esxi 4 host and I need to use fence_vmware_ng as a
> fence device. All works ok except when ESXi starts or is rebooted. I
> have configured under ESXi host to start automatically el5prodnode01
> only when host is rebooted or starts, but when el5prodnode01 guest
> automatically starts tries to launch el5prodnode02 every time. Why? Is
> this the normal procedure for fence_vmware_ng device?? How can I stop
> this feature or malfunction??
> 
> My cluster.conf is:
> 
> <?xml version="1.0"?>
> <cluster alias="VirtualRHELCluster" config_version="4" name="VirtCluster">
>         <fence_daemon post_fail_delay="0" post_join_delay="3"/>
>         <quorumd interval="1" tko="10" votes="1" label="prodqdisk"
> log_level="4">
>                 <heuristic program="ping 172.25.50.11 -c1 -t1" score="1"
> interval="2" tko="3"/>
>         </quorumd>
>         <clusternodes>
>                 <clusternode name="node01.hpulabs.org" nodeid="1"
> votes="1">
>                         <multicast addr="239.192.25.11" interface="eth2"/>
>                         <fence>
>                                 <method name="1">
>                                         <device name="vmware_fence_node1"/>
>                                 </method>
>                         </fence>
>                 </clusternode>
>                 <clusternode name="node02.hpulabs.org" nodeid="2"
> votes="1">
>                         <multicast addr="239.192.25.11" interface="eth2"/>
>                         <fence>
>                                 <method name="1">
>                                         <device name="vmware_fence_node2"/>
>                                 </method>
>                         </fence>
>                 </clusternode>
>         </clusternodes>
>         <cman expected_votes="3" two_node="0">
>                 <multicast addr="239.192.25.11"/>
>         </cman>
>         <fencedevices>
>                 <fencedevice agent="fence_vmware_ng"
> name="vmware_fence_node1" ipaddr="172.25.50.11" login="root"
> passwd="rootpass" port="el5prodnode01"/>
>                 <fencedevice agent="fence_vmware_ng"
> name="vmware_fence_node2" ipaddr="172.25.50.11" login="root"
> passwd="rootpass" port="el5prodnode02"/>
>                 <fencedevice agent="fence_manual" name="manual-fence"/>
>         </fencedevices>
>         <rm log_facility="local4" log_level="7">
>                 <failoverdomains>
>                         <failoverdomain name="PriCluster1" ordered="1"
> restricted="1">
>                                 <failoverdomainnode
> name="node01.hpulabs.org" priority="1"/>
>                                 <failoverdomainnode
> name="node02.hpulabs.org" priority="2"/>
>                         </failoverdomain>
>                         <failoverdomain name="PriCluster2" ordered="1"
> restricted="1">
>                                 <failoverdomainnode
> name="node02.hpulabs.org" priority="1"/>
>                                 <failoverdomainnode
> name="node01.hpulabs.org" priority="2"/>
>                         </failoverdomain>
>                         <failoverdomain name="FirstNode" restricted="1">
>                                 <failoverdomainnode
> name="node01.hpulabs.org" priority="1"/>
>                         </failoverdomain>
>                         <failoverdomain name="SecondNode" restricted="1">
>                                 <failoverdomainnode
> name="node02.hpulabs.org" priority="1"/>
>                         </failoverdomain>
>                 </failoverdomains>
>                 <resources>
>                         <fs device="/dev/clustervol/infravol"
> force_fsck="0" force_unmount="1" fstype="ext3"
> mountpoint="/data/services/infra" name="infradata" options="rw"/>
>                         <fs device="/dev/clustervol/mirrorvol"
> force_fsck="0" force_unmount="1" fstype="ext3"
> mountpoint="/data/services/www" name="mirrordata" options="rw"/>
>                         <script
> file="/data/config/etc/init.d/postfix-cluster" name="postfix-cluster"/>
>                 </resources>
>     </rm>
> </cluster>
> 
> Many thanks for your help.
> 

--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to