Looks like use_virsh=0 has no effect. -- [root@cvtst3 ~]# export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" ; OCF_RESKEY_use_virsh=0 [root@cvtst3 ~]# set -x ++ printf '\033]0;%s@%s:%s\007' root cvtst3 '~' [root@cvtst3 ~]# /usr/share/cluster/vm.sh start + /usr/share/cluster/vm.sh start Hypervisor: xen Management tool: virsh Hypervisor URI: xen:/// Migration URI format: xenmigr://target_host/ Virtual machine guest1 is error: failed to get domain 'guest1' error: Domain not found: xenUnifiedDomainLookupByName
<debug> virsh -c xen:/// start guest1 error: failed to get domain 'guest1' error: Domain not found: xenUnifiedDomainLookupByName ++ printf '\033]0;%s@%s:%s\007' root cvtst3 '~' [root@cvtst3 ~]# set +x + set +x --- -Paras. On Fri, Nov 22, 2013 at 5:22 PM, emmanuel segura <emi2f...@gmail.com> wrote: > Hellos Paras > > Stop the vm and retry to start the vm with following commands and if you > got some error show it > > export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" ; > OCF_RESKEY_use_virsh=0 > > > set -x > /usr/share/cluster/vm.sh start > set +x > > > 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> > >> I found the workaround to my issue. What i did is: >> >> run the vm using xm and then start using cluvscadm. This works for me for >> the time being but I am not sure what is causing this. This is what I did >> >> xm create /vms_c/guest1 >> clusvcadm -e vm: guest1 ( This detects that guest1 is up and quickly >> changes its status to success) >> >> Although i used virt-install, it also create a xem format configuration >> file and since use_virsh=0 it should be able to use this xen format config >> file. This is my vm configuration: >> >> --- >> name = "guest1" >> maxmem = 2048 >> memory = 512 >> vcpus = 1 >> #cpus="1-2" >> bootloader = "/usr/bin/pygrub" >> on_poweroff = "destroy" >> on_reboot = "restart" >> on_crash = "restart" >> vfb = [ ] >> disk = [ "tap:aio:/vms_c/guest1.img,xvda,w", >> "tap:aio:/vms_c/guest1-disk.img,xvdb,w" ] >> vif = [ "rate=10MB/s,mac=00:16:3e:6b:be:71,bridge=xenbr0" ] >> >> --- >> >> Thanks for you help Emmanuel ! Really appreciate it. >> >> -Paras. >> >> >> On Fri, Nov 22, 2013 at 11:10 AM, emmanuel segura <emi2f...@gmail.com>wrote: >> >>> ok, but your vm doesn't start on others nodes, i think, for >>> configuration problems >>> ================================================================ >>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on vm "guest1" >>> returned 1 (generic error) >>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: Failed to start >>> vm:guest1; return value: 1 >>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping service >>> vm:guest1 >>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is >>> recovering >>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: Relocating failed >>> service vm:guest1 >>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is >>> stopped >>> ================================================================ >>> in few words, try in every cluster node >>> >>> >>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>> >>> set -x >>> /usr/share/cluster/vm.sh start >>> /usr/share/cluster/vm.sh stop >>> >>> after you check if your vm can start and stop on every cluster node, >>> >>> /usr/share/cluster/vm.sh start >>> /usr/share/cluster/vm.sh migrate name_of_a_cluster_node >>> >>> can you show me your vm configuration under /vms_c? >>> >>> Thanks >>> Emmanuel >>> >>> >>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>> >>>> And also to test I made use_virsh=1 , same problem. The vm does not >>>> start up if the FD domains are offline. >>>> >>>> -Paras. >>>> >>>> >>>> On Fri, Nov 22, 2013 at 10:37 AM, Paras pradhan <pradhanpa...@gmail.com >>>> > wrote: >>>> >>>>> Well thats seems to theoretically correct. But right now my cluser has >>>>> use_virsh=0 and I don't have any issue untill my mebmers on the failover >>>>> domains are offline. So wondering what is it that clusvcadm -e is looking >>>>> when I don't use virsh . >>>>> >>>>> >>>>> >>>>> >>>>> On Fri, Nov 22, 2013 at 10:05 AM, emmanuel segura >>>>> <emi2f...@gmail.com>wrote: >>>>> >>>>>> If you used virt-install, i think you need use virsh, the cluster >>>>>> uses xm xen command if you got use_virsh=0 and virsh if you got >>>>>> use_virsh=1 >>>>>> in your cluster config >>>>>> >>>>>> >>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>> >>>>>>> I use virt-install to create virtual machines. Is there a way to >>>>>>> debug why clusvcadm -e vm:guest1 is failing? vm.sh seems to use virsh >>>>>>> and >>>>>>> my cluster.conf has use_virsh=0 >>>>>>> >>>>>>> >>>>>>> Thanks >>>>>>> >>>>>>> Paras. >>>>>>> >>>>>>> >>>>>>> On Nov 21, 2013 5:53 PM, "emmanuel segura" <emi2f...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> but did you configure your vm with xen tools or using virt-manager? >>>>>>>> >>>>>>>> >>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>> >>>>>>>>> Well no i don't want to use virsh. But as we are debugging with >>>>>>>>> virsh now i found a strange issue. >>>>>>>>> >>>>>>>>> I exported an xml file and imported to all nodes . Ran >>>>>>>>> >>>>>>>>> >>>>>>>>> --- >>>>>>>>> name="guest1" path="/vms_c" >>>>>>>>> >>>>>>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>>>>>> >>>>>>>>> set -x >>>>>>>>> /usr/share/cluster/vm.sh start >>>>>>>>> set +x >>>>>>>>> >>>>>>>>> -- >>>>>>>>> vm starts now. BUT from a cluster service : cluvscam -e vm:guest1 >>>>>>>>> , same error. >>>>>>>>> >>>>>>>>> >>>>>>>>> So if i populate all my domains' config files to all my cluser >>>>>>>>> nodes and make use_virsh=1, then the issue is resolved. But this is a >>>>>>>>> lot >>>>>>>>> of work for those who have hundreds of vm. >>>>>>>>> >>>>>>>>> vm.start uses virsh . Is there a way to tell him not use virsh? >>>>>>>>> >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> Paras. >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Nov 21, 2013 at 5:19 PM, emmanuel segura < >>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> if you are using virsh for manage your vms, change this in your >>>>>>>>>> cluster.conf >>>>>>>>>> >>>>>>>>>> from >>>>>>>>>> use_virsh="0" >>>>>>>>>> to >>>>>>>>>> use_virsh="1" >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>> >>>>>>>>>>> I think i found the problem. >>>>>>>>>>> >>>>>>>>>>> virsh list --all does not show my vm . This is because it was >>>>>>>>>>> created on another node. and another node has it. Now I want to >>>>>>>>>>> start the >>>>>>>>>>> service on a different node in which it was not created or where >>>>>>>>>>> virsh list >>>>>>>>>>> --all does not have an entry. Is it possible to create this entry >>>>>>>>>>> using a >>>>>>>>>>> xen config file?Looks like this is now a Xen issue rather than a >>>>>>>>>>> linux-cluster issue . :) >>>>>>>>>>> >>>>>>>>>>> Paras. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Nov 21, 2013 at 4:58 PM, emmanuel segura < >>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> 1:did you verify your xen livemigration configuration? >>>>>>>>>>>> 2: where you vm disk reside? >>>>>>>>>>>> 3: can you see your vm defined on every cluster node with xm >>>>>>>>>>>> list? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>> >>>>>>>>>>>>> This is what I get >>>>>>>>>>>>> >>>>>>>>>>>>> Hypervisor: xen >>>>>>>>>>>>> Management tool: virsh >>>>>>>>>>>>> Hypervisor URI: xen:/// >>>>>>>>>>>>> Migration URI format: xenmigr://target_host/ >>>>>>>>>>>>> Virtual machine guest1 is error: failed to get domain 'guest1' >>>>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>>>> >>>>>>>>>>>>> <debug> virsh -c xen:/// start guest1 >>>>>>>>>>>>> error: failed to get domain 'guest1' >>>>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>>>> >>>>>>>>>>>>> ++ printf '\033]0;%s@%s:%s\007' root vtst3 '~' >>>>>>>>>>>>> [root@cvtst3 ~]# set +x >>>>>>>>>>>>> + set +x >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> I am wondering why it failed to get domain . >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -Paras. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:43 PM, emmanuel segura < >>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> yes >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Well it is guest1. Isn't it?. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" >>>>>>>>>>>>>>> max_restarts="0" name="guest1" path="/vms_c" recovery="restart" >>>>>>>>>>>>>>> restart_expire_time="0" use_virsh="0"/> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> It is a vm service if it matters. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:22 PM, emmanuel segura < >>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> use the servicename you defined in your cluster.conf >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Says: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Running in test mode. >>>>>>>>>>>>>>>>> No resource guest1 of type service found >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:07 PM, emmanuel segura < >>>>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> rg_test test /etc/cluster/cluster.conf start service >>>>>>>>>>>>>>>>>> guest1 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> My failover domain looks like this: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="1" ordered="1" >>>>>>>>>>>>>>>>>>> restricted="0"> >>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>> name="vtst1" priority="1"/> >>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>> name="vtst3" priority="2"/> >>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>> name="vtst2" priority="3"/> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> </failoverdomain> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I have vm service that uses this failover domain. If my >>>>>>>>>>>>>>>>>>> node vtst1 is offline, the service doesnot start on vtst3 >>>>>>>>>>>>>>>>>>> which is 2nd in >>>>>>>>>>>>>>>>>>> the priority. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I tried to start it with: clusvcadm -e vm:guest1 and >>>>>>>>>>>>>>>>>>> even with -F and -m option. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> All i see is this error: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start >>>>>>>>>>>>>>>>>>> on vm "guest1" returned 1 (generic error) >>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: >>>>>>>>>>>>>>>>>>> Failed to start vm:guest1; return value: 1 >>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>> Stopping service vm:guest1 >>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service >>>>>>>>>>>>>>>>>>> vm:guest1 is recovering >>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: >>>>>>>>>>>>>>>>>>> Relocating failed service vm:guest1 >>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service >>>>>>>>>>>>>>>>>>> vm:guest1 is stopped >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> How do I debug? >>>>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > -- > Linux-cluster mailing list > Linux-cluster@redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster >
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster