I found the workaround to my issue. What i did is: run the vm using xm and then start using cluvscadm. This works for me for the time being but I am not sure what is causing this. This is what I did
xm create /vms_c/guest1 clusvcadm -e vm: guest1 ( This detects that guest1 is up and quickly changes its status to success) Although i used virt-install, it also create a xem format configuration file and since use_virsh=0 it should be able to use this xen format config file. This is my vm configuration: --- name = "guest1" maxmem = 2048 memory = 512 vcpus = 1 #cpus="1-2" bootloader = "/usr/bin/pygrub" on_poweroff = "destroy" on_reboot = "restart" on_crash = "restart" vfb = [ ] disk = [ "tap:aio:/vms_c/guest1.img,xvda,w", "tap:aio:/vms_c/guest1-disk.img,xvdb,w" ] vif = [ "rate=10MB/s,mac=00:16:3e:6b:be:71,bridge=xenbr0" ] --- Thanks for you help Emmanuel ! Really appreciate it. -Paras. On Fri, Nov 22, 2013 at 11:10 AM, emmanuel segura <emi2f...@gmail.com>wrote: > ok, but your vm doesn't start on others nodes, i think, for configuration > problems > ================================================================ > Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on vm "guest1" > returned 1 (generic error) > Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: Failed to start > vm:guest1; return value: 1 > Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping service > vm:guest1 > Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is > recovering > Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: Relocating failed > service vm:guest1 > Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is > stopped > ================================================================ > in few words, try in every cluster node > > > export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" > > set -x > /usr/share/cluster/vm.sh start > /usr/share/cluster/vm.sh stop > > after you check if your vm can start and stop on every cluster node, > > /usr/share/cluster/vm.sh start > /usr/share/cluster/vm.sh migrate name_of_a_cluster_node > > can you show me your vm configuration under /vms_c? > > Thanks > Emmanuel > > > 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> > >> And also to test I made use_virsh=1 , same problem. The vm does not start >> up if the FD domains are offline. >> >> -Paras. >> >> >> On Fri, Nov 22, 2013 at 10:37 AM, Paras pradhan >> <pradhanpa...@gmail.com>wrote: >> >>> Well thats seems to theoretically correct. But right now my cluser has >>> use_virsh=0 and I don't have any issue untill my mebmers on the failover >>> domains are offline. So wondering what is it that clusvcadm -e is looking >>> when I don't use virsh . >>> >>> >>> >>> >>> On Fri, Nov 22, 2013 at 10:05 AM, emmanuel segura <emi2f...@gmail.com>wrote: >>> >>>> If you used virt-install, i think you need use virsh, the cluster uses >>>> xm xen command if you got use_virsh=0 and virsh if you got use_virsh=1 in >>>> your cluster config >>>> >>>> >>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>> >>>>> I use virt-install to create virtual machines. Is there a way to debug >>>>> why clusvcadm -e vm:guest1 is failing? vm.sh seems to use virsh and my >>>>> cluster.conf has use_virsh=0 >>>>> >>>>> >>>>> Thanks >>>>> >>>>> Paras. >>>>> >>>>> >>>>> On Nov 21, 2013 5:53 PM, "emmanuel segura" <emi2f...@gmail.com> wrote: >>>>> >>>>>> but did you configure your vm with xen tools or using virt-manager? >>>>>> >>>>>> >>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>> >>>>>>> Well no i don't want to use virsh. But as we are debugging with >>>>>>> virsh now i found a strange issue. >>>>>>> >>>>>>> I exported an xml file and imported to all nodes . Ran >>>>>>> >>>>>>> >>>>>>> --- >>>>>>> name="guest1" path="/vms_c" >>>>>>> >>>>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>>>> >>>>>>> set -x >>>>>>> /usr/share/cluster/vm.sh start >>>>>>> set +x >>>>>>> >>>>>>> -- >>>>>>> vm starts now. BUT from a cluster service : cluvscam -e vm:guest1 , >>>>>>> same error. >>>>>>> >>>>>>> >>>>>>> So if i populate all my domains' config files to all my cluser nodes >>>>>>> and make use_virsh=1, then the issue is resolved. But this is a lot of >>>>>>> work >>>>>>> for those who have hundreds of vm. >>>>>>> >>>>>>> vm.start uses virsh . Is there a way to tell him not use virsh? >>>>>>> >>>>>>> >>>>>>> Thanks >>>>>>> Paras. >>>>>>> >>>>>>> >>>>>>> On Thu, Nov 21, 2013 at 5:19 PM, emmanuel segura <emi2f...@gmail.com >>>>>>> > wrote: >>>>>>> >>>>>>>> if you are using virsh for manage your vms, change this in your >>>>>>>> cluster.conf >>>>>>>> >>>>>>>> from >>>>>>>> use_virsh="0" >>>>>>>> to >>>>>>>> use_virsh="1" >>>>>>>> >>>>>>>> >>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>> >>>>>>>>> I think i found the problem. >>>>>>>>> >>>>>>>>> virsh list --all does not show my vm . This is because it was >>>>>>>>> created on another node. and another node has it. Now I want to start >>>>>>>>> the >>>>>>>>> service on a different node in which it was not created or where >>>>>>>>> virsh list >>>>>>>>> --all does not have an entry. Is it possible to create this entry >>>>>>>>> using a >>>>>>>>> xen config file?Looks like this is now a Xen issue rather than a >>>>>>>>> linux-cluster issue . :) >>>>>>>>> >>>>>>>>> Paras. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Nov 21, 2013 at 4:58 PM, emmanuel segura < >>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> 1:did you verify your xen livemigration configuration? >>>>>>>>>> 2: where you vm disk reside? >>>>>>>>>> 3: can you see your vm defined on every cluster node with xm list? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>> >>>>>>>>>>> This is what I get >>>>>>>>>>> >>>>>>>>>>> Hypervisor: xen >>>>>>>>>>> Management tool: virsh >>>>>>>>>>> Hypervisor URI: xen:/// >>>>>>>>>>> Migration URI format: xenmigr://target_host/ >>>>>>>>>>> Virtual machine guest1 is error: failed to get domain 'guest1' >>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>> >>>>>>>>>>> <debug> virsh -c xen:/// start guest1 >>>>>>>>>>> error: failed to get domain 'guest1' >>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>> >>>>>>>>>>> ++ printf '\033]0;%s@%s:%s\007' root vtst3 '~' >>>>>>>>>>> [root@cvtst3 ~]# set +x >>>>>>>>>>> + set +x >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> I am wondering why it failed to get domain . >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -Paras. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Nov 21, 2013 at 4:43 PM, emmanuel segura < >>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> yes >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>> >>>>>>>>>>>>> Well it is guest1. Isn't it?. >>>>>>>>>>>>> >>>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" >>>>>>>>>>>>> max_restarts="0" name="guest1" path="/vms_c" recovery="restart" >>>>>>>>>>>>> restart_expire_time="0" use_virsh="0"/> >>>>>>>>>>>>> >>>>>>>>>>>>> It is a vm service if it matters. >>>>>>>>>>>>> >>>>>>>>>>>>> -Paras. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:22 PM, emmanuel segura < >>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> use the servicename you defined in your cluster.conf >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Says: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Running in test mode. >>>>>>>>>>>>>>> No resource guest1 of type service found >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:07 PM, emmanuel segura < >>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> rg_test test /etc/cluster/cluster.conf start service guest1 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> My failover domain looks like this: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="1" ordered="1" >>>>>>>>>>>>>>>>> restricted="0"> >>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>> name="vtst1" priority="1"/> >>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>> name="vtst3" priority="2"/> >>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>> name="vtst2" priority="3"/> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> </failoverdomain> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I have vm service that uses this failover domain. If my >>>>>>>>>>>>>>>>> node vtst1 is offline, the service doesnot start on vtst3 >>>>>>>>>>>>>>>>> which is 2nd in >>>>>>>>>>>>>>>>> the priority. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I tried to start it with: clusvcadm -e vm:guest1 and >>>>>>>>>>>>>>>>> even with -F and -m option. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> All i see is this error: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on >>>>>>>>>>>>>>>>> vm "guest1" returned 1 (generic error) >>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: >>>>>>>>>>>>>>>>> Failed to start vm:guest1; return value: 1 >>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping >>>>>>>>>>>>>>>>> service vm:guest1 >>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service >>>>>>>>>>>>>>>>> vm:guest1 is recovering >>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: >>>>>>>>>>>>>>>>> Relocating failed service vm:guest1 >>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service >>>>>>>>>>>>>>>>> vm:guest1 is stopped >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> How do I debug? >>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster@redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster >
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster