Yes strange.

cluster.conf 's entry is

<vm autostart="0" domain="myfd1" exclusive="0" max_restarts="0"
name="guest1" path="/vms" recovery="restart" restart_expire_time="0"/>


Paras.


On Tue, Sep 29, 2009 at 4:02 PM, brem belguebli
<[email protected]> wrote:
> That looks strange, maybe your cluster.conf will help
>
> 2009/9/29 Paras pradhan <[email protected]>:
>> I don't see rgmanager .
>>
>> Here is the o/p from clustat
>>
>> [r...@cvtst1 cluster]# clustat
>> Cluster Status for test @ Tue Sep 29 15:53:33 2009
>> Member Status: Quorate
>>
>>  Member Name                                                     ID   Status
>>  ------ ----                                                     ---- ------
>>  cvtst2                                                    1 Online
>>  cvtst1                                                     2 Online, Local
>>  cvtst3                                                     3 Online
>>
>>
>> Thanks
>> Paras.
>>
>> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli
>> <[email protected]> wrote:
>>> It looks correct, rgmanager seems to start on all nodes
>>>
>>> what gives you clustat ?
>>>
>>> If rgmanager doesn't show, check out the logs something may have gone wrong.
>>>
>>>
>>> 2009/9/29 Paras pradhan <[email protected]>:
>>>> Change to 7 and i got this log
>>>>
>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down
>>>> Cluster Service Manager...
>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, 
>>>> exiting
>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service
>>>> Manager is stopped.
>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group
>>>> Manager Starting
>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data
>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP
>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP
>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed
>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed
>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed
>>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed
>>>>
>>>>
>>>> Anything unusual here?
>>>>
>>>> Paras.
>>>>
>>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
>>>> <[email protected]> wrote:
>>>>> I use log_level=7 to have more debugging info.
>>>>>
>>>>> It seems 4 is not enough.
>>>>>
>>>>> Brem
>>>>>
>>>>>
>>>>> 2009/9/29, Paras pradhan <[email protected]>:
>>>>>> Withe log_level of 3 I got only this
>>>>>>
>>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down
>>>>>> Cluster Service Manager...
>>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down
>>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, 
>>>>>> exiting
>>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service
>>>>>> Manager is stopped.
>>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group
>>>>>> Manager Starting
>>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down
>>>>>> Cluster Service Manager...
>>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service
>>>>>> Manager is stopped.
>>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group
>>>>>> Manager Starting
>>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down
>>>>>>
>>>>>> I do not know what the last line means.
>>>>>>
>>>>>> rgmanager version I am running is:
>>>>>> rgmanager-2.0.52-1.el5.centos
>>>>>>
>>>>>> I don't what has gone wrong.
>>>>>>
>>>>>> Thanks
>>>>>> Paras.
>>>>>>
>>>>>>
>>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
>>>>>> <[email protected]> wrote:
>>>>>> > you mean it stopped successfully on all the nodes but it is failing to
>>>>>> > start only on node cvtst1 ?
>>>>>> >
>>>>>> > look at the following page  to make rgmanager more verbose. It 'll
>>>>>> > help debug....
>>>>>> >
>>>>>> > http://sources.redhat.com/cluster/wiki/RGManager
>>>>>> >
>>>>>> > at Logging Configuration section
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > 2009/9/29 Paras pradhan <[email protected]>:
>>>>>> >> Brem,
>>>>>> >>
>>>>>> >> When I try to restart rgmanager on all the nodes, this time i do not
>>>>>> >> see rgmanager running on the first node. But I do see on other 2
>>>>>> >> nodes.
>>>>>> >>
>>>>>> >> Log on the first node:
>>>>>> >>
>>>>>> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group
>>>>>> >> Manager Starting
>>>>>> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down
>>>>>> >> Cluster Service Manager...
>>>>>> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down
>>>>>> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, 
>>>>>> >> exiting
>>>>>> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service
>>>>>> >> Manager is stopped.
>>>>>> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group
>>>>>> >> Manager Starting
>>>>>> >>
>>>>>> >> -
>>>>>> >> It seems service is running ,  but I do not see rgmanger running 
>>>>>> >> using clustat
>>>>>> >>
>>>>>> >>
>>>>>> >> Don't know what is going on.
>>>>>> >>
>>>>>> >> Thanks
>>>>>> >> Paras.
>>>>>> >>
>>>>>> >>
>>>>>> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
>>>>>> >> <[email protected]> wrote:
>>>>>> >>> Paras,
>>>>>> >>>
>>>>>> >>> Another thing, it would have been more interesting to have a start
>>>>>> >>> DEBUG not a stop.
>>>>>> >>>
>>>>>> >>> That's why I was asking you to first stop the vm manually on all your
>>>>>> >>> nodes, stop eventually rgmanager on all the nodes to reset the
>>>>>> >>> potential wrong states you may have, restart rgmanager.
>>>>>> >>>
>>>>>> >>> If your VM is configured to autostart, this will make it start.
>>>>>> >>>
>>>>>> >>> It should normally fail (as it does now). Send out your newly created
>>>>>> >>> DEBUG file.
>>>>>> >>>
>>>>>> >>> 2009/9/29 brem belguebli <[email protected]>:
>>>>>> >>>> Hi Paras,
>>>>>> >>>>
>>>>>> >>>>
>>>>>> >>>> I don't know the xen/cluster combination well, but if I do remember
>>>>>> >>>> well, I think I've read somewhere that when using xen you have to
>>>>>> >>>> declare the use_virsh=0 key in the VM definition in the 
>>>>>> >>>> cluster.conf.
>>>>>> >>>>
>>>>>> >>>> This would make rgmanager use xm commands instead of virsh
>>>>>> >>>> The DEBUG output shows clearly that you are using virsh to manage 
>>>>>> >>>> your
>>>>>> >>>> VM instead of xm commands.
>>>>>> >>>> Check out the RH docs about virtualization
>>>>>> >>>>
>>>>>> >>>> I'm not a 100% sure about that, I may be completely wrong.
>>>>>> >>>>
>>>>>> >>>> Brem
>>>>>> >>>>
>>>>>> >>>> 2009/9/28 Paras pradhan <[email protected]>:
>>>>>> >>>>> The only thing I noticed is the message after stopping the vm 
>>>>>> >>>>> using xm
>>>>>> >>>>> in all nodes and starting using clusvcadm is
>>>>>> >>>>>
>>>>>> >>>>> "Virtual machine guest1 is blocked"
>>>>>> >>>>>
>>>>>> >>>>> The whole DEBUG file is attached.
>>>>>> >>>>>
>>>>>> >>>>>
>>>>>> >>>>> Thanks
>>>>>> >>>>> Paras.
>>>>>> >>>>>
>>>>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
>>>>>> >>>>> <[email protected]> wrote:
>>>>>> >>>>>> There's a problem with the script that is called by rgmanager to 
>>>>>> >>>>>> start
>>>>>> >>>>>> the VM, I don't know what causes it
>>>>>> >>>>>>
>>>>>> >>>>>> May be you should try something like :
>>>>>> >>>>>>
>>>>>> >>>>>> 1) stop the VM on all nodes with xm commands
>>>>>> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following
>>>>>> >>>>>> lines (after the #!/bin/bash ):
>>>>>> >>>>>>   exec >/tmp/DEBUG 2>&1
>>>>>> >>>>>>   set -x
>>>>>> >>>>>> 3) start the VM with clusvcadm -e vm:guest1
>>>>>> >>>>>>
>>>>>> >>>>>> It should fail as it did before.
>>>>>> >>>>>>
>>>>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it
>>>>>> >>>>>> fails (it may generate a lot of debug)
>>>>>> >>>>>>
>>>>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh
>>>>>> >>>>>>
>>>>>> >>>>>> Post the DEBUG file if you're not able to see where it fails.
>>>>>> >>>>>>
>>>>>> >>>>>> Brem
>>>>>> >>>>>>
>>>>>> >>>>>> 2009/9/26 Paras pradhan <[email protected]>:
>>>>>> >>>>>>> No I am not manually starting not using automatic init scripts.
>>>>>> >>>>>>>
>>>>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1
>>>>>> >>>>>>>
>>>>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few 
>>>>>> >>>>>>> seconds it
>>>>>> >>>>>>> says guest1 started . But after a while I can see the guest1 on 
>>>>>> >>>>>>> all
>>>>>> >>>>>>> three nodes.
>>>>>> >>>>>>>
>>>>>> >>>>>>> clustat says:
>>>>>> >>>>>>>
>>>>>> >>>>>>>  Service Name                                            Owner 
>>>>>> >>>>>>> (Last)
>>>>>> >>>>>>>                                          State
>>>>>> >>>>>>>  ------- ----                                            ----- 
>>>>>> >>>>>>> ------
>>>>>> >>>>>>>                                          -----
>>>>>> >>>>>>>  vm:guest1                                               (none)
>>>>>> >>>>>>>                                          stopped
>>>>>> >>>>>>>
>>>>>> >>>>>>> But I can see the vm from xm li.
>>>>>> >>>>>>>
>>>>>> >>>>>>> This is what I can see from the log:
>>>>>> >>>>>>>
>>>>>> >>>>>>>
>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm 
>>>>>> >>>>>>> "guest1"
>>>>>> >>>>>>> returned 1 (generic error)
>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to 
>>>>>> >>>>>>> start
>>>>>> >>>>>>> vm:guest1; return value: 1
>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping 
>>>>>> >>>>>>> service vm:guest1
>>>>>> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service 
>>>>>> >>>>>>> vm:guest1 is
>>>>>> >>>>>>> recovering
>>>>>> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering 
>>>>>> >>>>>>> failed
>>>>>> >>>>>>> service vm:guest1
>>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm 
>>>>>> >>>>>>> "guest1"
>>>>>> >>>>>>> returned 1 (generic error)
>>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to 
>>>>>> >>>>>>> start
>>>>>> >>>>>>> vm:guest1; return value: 1
>>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping 
>>>>>> >>>>>>> service vm:guest1
>>>>>> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service 
>>>>>> >>>>>>> vm:guest1 is
>>>>>> >>>>>>> recovering
>>>>>> >>>>>>>
>>>>>> >>>>>>>
>>>>>> >>>>>>> Paras.
>>>>>> >>>>>>>
>>>>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>>>>>> >>>>>>> <[email protected]> wrote:
>>>>>> >>>>>>>> Have you started  your VM via rgmanager (clusvcadm -e 
>>>>>> >>>>>>>> vm:guest1) or
>>>>>> >>>>>>>> using xm commands out of cluster control  (or maybe a thru an
>>>>>> >>>>>>>> automatic init script ?)
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> When clustered, you should never be starting services (manually 
>>>>>> >>>>>>>> or
>>>>>> >>>>>>>> thru automatic init script) out of cluster control
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the 
>>>>>> >>>>>>>> adequate
>>>>>> >>>>>>>> xm command (not using xen myself) and try to start it with 
>>>>>> >>>>>>>> clusvcadm.
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> Then see if it is started on all nodes (send clustat output)
>>>>>> >>>>>>>>
>>>>>> >>>>>>>>
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> 2009/9/25 Paras pradhan <[email protected]>:
>>>>>> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though 
>>>>>> >>>>>>>>> clustat
>>>>>> >>>>>>>>> says it is stopped.
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> --
>>>>>> >>>>>>>>> [r...@cvtst1 ~]# clustat
>>>>>> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>>>>> >>>>>>>>> Member Status: Quorate
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>>  Member Name                                                   
>>>>>> >>>>>>>>>   ID   Status
>>>>>> >>>>>>>>>  ------ ----                                                   
>>>>>> >>>>>>>>>   ---- ------
>>>>>> >>>>>>>>>  cvtst2                                                    1 
>>>>>> >>>>>>>>> Online, rgmanager
>>>>>> >>>>>>>>>  cvtst1                                                     2 
>>>>>> >>>>>>>>> Online,
>>>>>> >>>>>>>>> Local, rgmanager
>>>>>> >>>>>>>>>  cvtst3                                                     3 
>>>>>> >>>>>>>>> Online, rgmanager
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>>  Service Name                                            Owner 
>>>>>> >>>>>>>>> (Last)
>>>>>> >>>>>>>>>                                          State
>>>>>> >>>>>>>>>  ------- ----                                            ----- 
>>>>>> >>>>>>>>> ------
>>>>>> >>>>>>>>>                                          -----
>>>>>> >>>>>>>>>  vm:guest1                                               (none)
>>>>>> >>>>>>>>>                                          stopped
>>>>>> >>>>>>>>> [r...@cvtst1 ~]#
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> ---
>>>>>> >>>>>>>>> o/p of xm li on cvtst1
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> --
>>>>>> >>>>>>>>> [r...@cvtst1 ~]# xm li
>>>>>> >>>>>>>>> Name                                      ID Mem(MiB) VCPUs 
>>>>>> >>>>>>>>> State   Time(s)
>>>>>> >>>>>>>>> Domain-0                                   0     3470     2 
>>>>>> >>>>>>>>> r-----  28939.4
>>>>>> >>>>>>>>> guest1                                     7      511     1 
>>>>>> >>>>>>>>> -b----   7727.8
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> o/p of xm li on cvtst2
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> --
>>>>>> >>>>>>>>> [r...@cvtst2 ~]# xm li
>>>>>> >>>>>>>>> Name                                      ID Mem(MiB) VCPUs 
>>>>>> >>>>>>>>> State   Time(s)
>>>>>> >>>>>>>>> Domain-0                                   0     3470     2 
>>>>>> >>>>>>>>> r-----  31558.9
>>>>>> >>>>>>>>> guest1                                    21      511     1 
>>>>>> >>>>>>>>> -b----   7558.2
>>>>>> >>>>>>>>> ---
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> Thanks
>>>>>> >>>>>>>>> Paras.
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>>>>> >>>>>>>>> <[email protected]> wrote:
>>>>>> >>>>>>>>>> It looks like no.
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> can you send an output of clustat  of when the VM is running 
>>>>>> >>>>>>>>>> on
>>>>>> >>>>>>>>>> multiple nodes at the same time?
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm 
>>>>>> >>>>>>>>>> -s vm:guest1) ?
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> 2009/9/25 Paras pradhan <[email protected]>:
>>>>>> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not 
>>>>>> >>>>>>>>>>> being
>>>>>> >>>>>>>>>>> properly handled by the cluster.
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> Thanks
>>>>>> >>>>>>>>>>> Paras.
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan 
>>>>>> >>>>>>>>>>> <[email protected]> wrote:
>>>>>> >>>>>>>>>>>> Ok.. here is my cluster.conf file
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> --
>>>>>> >>>>>>>>>>>> [r...@cvtst1 cluster]# more cluster.conf
>>>>>> >>>>>>>>>>>> <?xml version="1.0"?>
>>>>>> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test">
>>>>>> >>>>>>>>>>>>        <fence_daemon clean_start="0" post_fail_delay="0" 
>>>>>> >>>>>>>>>>>> post_join_delay="3"/>
>>>>>> >>>>>>>>>>>>        <clusternodes>
>>>>>> >>>>>>>>>>>>                <clusternode name="cvtst2" nodeid="1" 
>>>>>> >>>>>>>>>>>> votes="1">
>>>>>> >>>>>>>>>>>>                        <fence/>
>>>>>> >>>>>>>>>>>>                </clusternode>
>>>>>> >>>>>>>>>>>>                <clusternode name="cvtst1" nodeid="2" 
>>>>>> >>>>>>>>>>>> votes="1">
>>>>>> >>>>>>>>>>>>                        <fence/>
>>>>>> >>>>>>>>>>>>                </clusternode>
>>>>>> >>>>>>>>>>>>                <clusternode name="cvtst3" nodeid="3" 
>>>>>> >>>>>>>>>>>> votes="1">
>>>>>> >>>>>>>>>>>>                        <fence/>
>>>>>> >>>>>>>>>>>>                </clusternode>
>>>>>> >>>>>>>>>>>>        </clusternodes>
>>>>>> >>>>>>>>>>>>        <cman/>
>>>>>> >>>>>>>>>>>>        <fencedevices/>
>>>>>> >>>>>>>>>>>>        <rm>
>>>>>> >>>>>>>>>>>>                <failoverdomains>
>>>>>> >>>>>>>>>>>>                        <failoverdomain name="myfd1" 
>>>>>> >>>>>>>>>>>> nofailback="0" ordered="1" restricted="0">
>>>>>> >>>>>>>>>>>>                                <failoverdomainnode 
>>>>>> >>>>>>>>>>>> name="cvtst2" priority="3"/>
>>>>>> >>>>>>>>>>>>                                <failoverdomainnode 
>>>>>> >>>>>>>>>>>> name="cvtst1" priority="1"/>
>>>>>> >>>>>>>>>>>>                                <failoverdomainnode 
>>>>>> >>>>>>>>>>>> name="cvtst3" priority="2"/>
>>>>>> >>>>>>>>>>>>                        </failoverdomain>
>>>>>> >>>>>>>>>>>>                </failoverdomains>
>>>>>> >>>>>>>>>>>>                <resources/>
>>>>>> >>>>>>>>>>>>                <vm autostart="1" domain="myfd1" 
>>>>>> >>>>>>>>>>>> exclusive="0" max_restarts="0"
>>>>>> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r
>>>>>> >>>>>>>>>>>> estart" restart_expire_time="0"/>
>>>>>> >>>>>>>>>>>>        </rm>
>>>>>> >>>>>>>>>>>> </cluster>
>>>>>> >>>>>>>>>>>> [r...@cvtst1 cluster]#
>>>>>> >>>>>>>>>>>> ------
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> Thanks!
>>>>>> >>>>>>>>>>>> Paras.
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer 
>>>>>> >>>>>>>>>>>> <[email protected]> wrote:
>>>>>> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>>>>> >>>>>>>>>>>>> Paras pradhan <[email protected]> wrote:
>>>>>> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. 
>>>>>> >>>>>>>>>>>>>> Now I am
>>>>>> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one 
>>>>>> >>>>>>>>>>>>>> node, it
>>>>>> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls  this?
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the 
>>>>>> >>>>>>>>>>>>> rgmanager
>>>>>> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. 
>>>>>> >>>>>>>>>>>>> Maybe,
>>>>>> >>>>>>>>>>>>> you can post your cluster.conf?
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> Regards,
>>>>>> >>>>>>>>>>>>> Volker
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> --
>>>>>> >>>>>>>>>>>>> Linux-cluster mailing list
>>>>>> >>>>>>>>>>>>> [email protected]
>>>>>> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> --
>>>>>> >>>>>>>>>>> Linux-cluster mailing list
>>>>>> >>>>>>>>>>> [email protected]
>>>>>> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> --
>>>>>> >>>>>>>>>> Linux-cluster mailing list
>>>>>> >>>>>>>>>> [email protected]
>>>>>> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> --
>>>>>> >>>>>>>>> Linux-cluster mailing list
>>>>>> >>>>>>>>> [email protected]
>>>>>> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> --
>>>>>> >>>>>>>> Linux-cluster mailing list
>>>>>> >>>>>>>> [email protected]
>>>>>> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>>>
>>>>>> >>>>>>>
>>>>>> >>>>>>> --
>>>>>> >>>>>>> Linux-cluster mailing list
>>>>>> >>>>>>> [email protected]
>>>>>> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>>
>>>>>> >>>>>>
>>>>>> >>>>>> --
>>>>>> >>>>>> Linux-cluster mailing list
>>>>>> >>>>>> [email protected]
>>>>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>>
>>>>>> >>>>>
>>>>>> >>>>> --
>>>>>> >>>>> Linux-cluster mailing list
>>>>>> >>>>> [email protected]
>>>>>> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>>>
>>>>>> >>>>
>>>>>> >>>
>>>>>> >>> --
>>>>>> >>> Linux-cluster mailing list
>>>>>> >>> [email protected]
>>>>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>>
>>>>>> >>
>>>>>> >> --
>>>>>> >> Linux-cluster mailing list
>>>>>> >> [email protected]
>>>>>> >> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >>
>>>>>> >
>>>>>> > --
>>>>>> > Linux-cluster mailing list
>>>>>> > [email protected]
>>>>>> > https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>> >
>>>>>>
>>>>>> --
>>>>>> Linux-cluster mailing list
>>>>>> [email protected]
>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>
>>>>>
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> [email protected]
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>
>>>>
>>>> --
>>>> Linux-cluster mailing list
>>>> [email protected]
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>
>>>
>>> --
>>> Linux-cluster mailing list
>>> [email protected]
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>
>> --
>> Linux-cluster mailing list
>> [email protected]
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
> --
> Linux-cluster mailing list
> [email protected]
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

--
Linux-cluster mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to