Re: [Users] ovirt shows vm status as down, but vm is actually up and running
You can try to compress the file: sudo yum install bzip2 bzip2 vdsm.log On Fri, May 3, 2013 at 10:08 AM, Pradipta Kumar Banerjee wrote: > On 05/03/2013 01:10 AM, Dan Kenigsberg wrote: > > On Thu, May 02, 2013 at 11:26:41PM +0530, Pradipta Kumar Banerjee wrote: > >> Hi, > >> I have a situation where VM is actually up and running (I can access > its VNC > >> console), however oVirt engine ui shows the status as down. > >> I'm using ovirt engine v3.2 from stable repo on F18 > >> Any suggestions to trouble shoot the issue ? > > > > That's like the worse thing that can happen - if you ever make decide to > > start another instance of that VM on another host - so please tread > > carefully. > > > > Does the host that run the VM shows as Up in Engine? > Yes the host is up > > What does > > vdsClient -s 0 list table > > shows there? > > Empty output > > > > and > > virsh -r list > This lists the VM. Also I'm able to access the console of the VM and work > on it. > > > ? > > > > When is the last time that the VM was known to be up by Engine? Do you > > have vdsm.log from that time? > > > > > It is very important to the project that you collect as much information > > about the lifecycle of this VM (i.e. was it migrated from somewhere, > > complete vdsm.log since creation). However, you may have a conflicting > > interest if you value that VM's data. You should probably shut it down > > cleanly from within the guest in that case. > No migration, VM pinned to specific host itself. > Its a test setup so no worries on data corruption. However I agree its > important > to find out the root cause. > > Btw I should mention that I'm using ovirt-engine on x86_64 and vdsm on IBM > Power > host. Not sure if its a host specific issue. > I'll see if I can recreate the issue. > > vdsm.log file is around 11MB. Is there any place where I can upload it? > > > > > > > Dan. > > > > > -- > Regards, > Pradipta > > ___ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Unable to add host in a development setup - "Host kernelpanic.home installation failed. Failed to decode own public key from store 'keys/engine.p12' using alias '1'."
Thank you very much for your prompt response. However, according to "git rebase" (after "git fetch") there is no new change [wil@bufferoverflow repository]$ git fetch https://github.com/alonbl/ovirt-engine.git otopi >From https://github.com/alonbl/ovirt-engine * branchotopi -> FETCH_HEAD [wil@bufferoverflow repository]$ git rebase --merge FETCH_HEAD *Current branch otopi is up to date.* * * The latest change I see in "git log" is [wil@bufferoverflow repository]$ git show HEAD commit 99d95891ab0e814159a463d246543b08e41c746d Author: Alon Bar-Lev Date: Thu Apr 25 12:39:50 2013 +0300 packaging: engine-setup re-write using otopi Change-Id: I4d5e3ac7927802326e324c9b98ada75b421638aa Signed-off-by: Alon Bar-Lev Signed-off-by: Sandro Bonazzola Signed-off-by: Alex Lourie If you don't wish to push the change yet, you can send a patch and I will apply it. On Mon, Apr 29, 2013 at 1:04 AM, Alon Bar-Lev wrote: > > > ----- Original Message - > > From: "Limor Gavish" > > To: "Alon Bar-Lev" > > Cc: "Yuval M" , users@ovirt.org, "Nezer Zaidenberg" < > nzaidenb...@mac.com>, "Vered Volansky" > > > > Sent: Sunday, April 28, 2013 10:56:45 PM > > Subject: Re: [Users] Unable to add host in a development setup - "Host > kernelpanic.home installation failed. Failed > > to decode own public key from store 'keys/engine.p12' using alias '1'." > > > > Hi, > > > > Thank you very much for your help but I do not succeed to start the > engine > > after compiling your version. > > > > I switched to the otopi branch like you said. > > git fetch https://github.com/alonbl/ovirt-engine.git otopi > > git checkout -b otopi FETCH_HEAD > > > > I followed the instructions in > > https://github.com/alonbl/ovirt-engine/blob/otopi/README.developer > > *So here are the last 3 commands I used:* > > make clean install-dev PREFIX=$HOME/ovirt-engine > > $HOME/ovirt-engine/bin/engine-setup-2 > > --jboss-home=$HOME/ovirt-engine/jboss-as-7.1.1.Final > > $HOME/ovirt-engine/share/ovirt-engine/services/ovirt-engine.py start > > > > However, the last command > > ($HOME/ovirt-engine/share/ovirt-engine/services/ovirt-engine.py start) > > hangs. > > This is expected... it is the service running, Ctrl-C to terminate. > > > *In server.log I see the following error:* > > 2013-04-28 22:13:40,203 ERROR [org.jboss.as] (MSC service thread 1-3) > > JBAS015875: JBoss AS 7.1.1.Final "Brontes" started (with errors) in > 2788ms > > - Started 504 of 5 40 services (1 services failed or missing > > dependencies, 34 services are passive or on-demand) > > *In engine.log I see the following error:* > > 2013-04-28 22:13:40,082 ERROR > > [org.ovirt.engine.core.utils.ErrorTranslatorImpl] (MSC service thread > 1-4) > > File: bundles/VdsmErrors could not be loaded: java.util.Mis > > singResourceException: Can't find bundle for base name > bundles/VdsmErrors, > > locale en_US > > > > Full logs are attached. > > The problem is that for some reason the PREFIX is not taken into account. > > Related errors: > 2013-04-28 19:57:46,468 WARN [org.ovirt.engine.core.utils.LocalConfig] > (MSC service thread 1-3) The file > "/usr/share/ovirt-engine/conf/engine.conf.defaults" doesn't exist or isn't > readable. Will return an empty set of properties. > 2013-04-28 19:57:46,468 WARN [org.ovirt.engine.core.utils.LocalConfig] > (MSC service thread 1-3) The file "/etc/ovirt-engine/engine.conf" doesn't > exist or isn't readable. Will return an empty set of properties. > > It can be due to previous make execution without PREFIX. > I've committed a change to Makefile to force regenerate the file locations > when using install-dev. > Can you please give it a try? > > Thanks! > > > Thanks, > > Limor on behalf of Yuval > > > > > > On Sun, Apr 28, 2013 at 2:51 PM, Alon Bar-Lev wrote: > > > > > > > > Maybe someone can help you in this. > > > > > > I prefer people install development mode and use fully functional > product > > > within their development environment. > > > > > > Alon > > > > > > - Original Message - > > > > From: "Yuval M" > > > > To: "Alon Bar-Lev" > > > > Cc: users@ovirt.org, "Nezer Zaidenberg" , > "Limor > > > Gavish" , "Vered Volansky" > > > > > > > > Sent: Sunday,
Re: [Users] oVirt storage is down and doesn't come up
r a Postgres ui if you have one. > In the results you will see the storage connection in the format of > %hostname%:/%mountName%, then in the VDSM server check in the mount list > that you see that it is mounted, the mount itself should contain a > directory named as the uuid of the master domain, let me know the result. > > Tal. > > > > > On 04/12/2013 07:29 PM, Limor Gavish wrote: > > Hi, > > For some reason, without doing anything, all the storage domains became > down and restarting VDSM or the entire machine do not bring it up. > I am not using lvm > The following errors appear several times in vdsm.log (full logs are > attached): > > Thread-22::WARNING::2013-04-12 > 19:00:08,597::lvm::378::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] [' > Volume group "1083422e-a5db-41b6-b667-b9ef1ef244f0" not found'] > Thread-22::DEBUG::2013-04-12 > 19:00:08,598::lvm::402::OperationMutex::(_reloadvgs) Operation 'lvm reload > operation' released the operation mutex > Thread-22::DEBUG::2013-04-12 > 19:00:08,681::resourceManager::615::ResourceManager::(releaseResource) > Trying to release resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' > Thread-22::DEBUG::2013-04-12 > 19:00:08,681::resourceManager::634::ResourceManager::(releaseResource) > Released resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' (0 active > users) > Thread-22::DEBUG::2013-04-12 > 19:00:08,681::resourceManager::640::ResourceManager::(releaseResource) > Resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' is free, finding > out if anyone is waiting for it. > Thread-22::DEBUG::2013-04-12 > 19:00:08,682::resourceManager::648::ResourceManager::(releaseResource) No > one is waiting for resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3', > Clearing records. > Thread-22::ERROR::2013-04-12 > 19:00:08,682::task::850::TaskManager.Task::(_setError) > Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::Unexpected error > Traceback (most recent call last): > File "/usr/share/vdsm/storage/task.py", line 857, in _run > return fn(*args, **kargs) > File "/usr/share/vdsm/logUtils.py", line 45, in wrapper > res = f(*args, **kwargs) > File "/usr/share/vdsm/storage/hsm.py", line 939, in connectStoragePool > masterVersion, options) > File "/usr/share/vdsm/storage/hsm.py", line 986, in _connectStoragePool > res = pool.connect(hostID, scsiKey, msdUUID, masterVersion) > File "/usr/share/vdsm/storage/sp.py", line 695, in connect > self.__rebuild(msdUUID=msdUUID, masterVersion=masterVersion) > File "/usr/share/vdsm/storage/sp.py", line 1232, in __rebuild > masterVersion=masterVersion) > File "/usr/share/vdsm/storage/sp.py", line 1576, in getMasterDomain > raise se.StoragePoolMasterNotFound(self.spUUID, msdUUID) > StoragePoolMasterNotFound: Cannot find master domain: > 'spUUID=5849b030-626e-47cb-ad90-3ce782d831b3, > msdUUID=1083422e-a5db-41b6-b667-b9ef1ef244f0' > Thread-22::DEBUG::2013-04-12 > 19:00:08,685::task::869::TaskManager.Task::(_run) > Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::Task._run: > e35a22ac-771a-4916-851f-2fe9d60a0ae6 > ('5849b030-626e-47cb-ad90-3ce782d831b3', 1, > '5849b030-626e-47cb-ad90-3ce782d831b3', > '1083422e-a5db-41b6-b667-b9ef1ef244f0', 3942) {} failed - stopping task > Thread-22::DEBUG::2013-04-12 > 19:00:08,685::task::1194::TaskManager.Task::(stop) > Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::stopping in state preparing > (force False) > Thread-22::DEBUG::2013-04-12 > 19:00:08,685::task::974::TaskManager.Task::(_decref) > Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::ref 1 aborting True > Thread-22::INFO::2013-04-12 > 19:00:08,686::task::1151::TaskManager.Task::(prepare) > Task=`e35a22ac-771a-4916-851f-2fe9d60a0ae6`::aborting: Task is aborted: > 'Cannot find master domain' - code 304 > > *[wil@bufferoverflow ~]$ **sudo vgs --noheadings --units b --nosuffix > --separator \| -o > uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free > * > No volume groups found > > *[wil@bufferoverflow ~]$ **mount* > proc on /proc type proc (rw,nosuid,nodev,noexec,relatime) > sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime) > devtmpfs on /dev type devtmpfs > (rw,nosuid,size=8131256k,nr_inodes=2032814,mode=755) > securityfs on /sys/kernel/security type securityfs > (rw,nosuid,nodev,noexec,relatime) > tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev) > devpts on /dev/pts type devpts > (rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000) > tmpfs on /run type tmpfs (rw,nosuid,nodev,
Re: [Users] VM crashes and doesn't recover
t;> svdsm >> > failed [Errno 2] No such file or directory >> > Mar 24 19:57:52 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to >> svdsm >> > failed [Errno 2] No such file or directory >> > Mar 24 19:57:53 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to >> svdsm >> > failed [Errno 2] No such file or directory >> > Mar 24 19:57:54 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to >> svdsm >> > failed [Errno 2] No such file or directory >> > Mar 24 19:57:55 bufferoverflow vdsm SuperVdsmProxy WARNING Connect to >> svdsm >> > failed [Errno 2] No such file or directory >> > Mar 24 19:57:55 bufferoverflow vdsm Storage.Misc ERROR Panic: Couldn't >> > connect to supervdsm >> > Mar 24 19:57:55 bufferoverflow respawn: slave '/usr/share/vdsm/vdsm' >> died, >> > respawning slave >> > Mar 24 19:57:55 bufferoverflow vdsm fileUtils WARNING Dir >> > /rhev/data-center/mnt already exists >> > Mar 24 19:57:58 bufferoverflow vdsm vds WARNING Unable to load the json >> rpc >> > server module. Please make sure it is installed. >> > Mar 24 19:57:58 bufferoverflow vdsm vm.Vm WARNING >> > vmId=`4d3d81b3-d083-4569-acc2-8e631ed51843`::Unknown type found, device: >> > '{'device': u'unix', 'alias': u'channel0', 'type': u'channel', >> 'address': >> > {u'bus': u'0', u'controller': u'0', u'type': u'virtio-serial', u'port': >> > u'1'}}' found >> > Mar 24 19:57:58 bufferoverflow vdsm vm.Vm WARNING >> > vmId=`4d3d81b3-d083-4569-acc2-8e631ed51843`::Unknown type found, device: >> > '{'device': u'unix', 'alias': u'channel1', 'type': u'channel', >> 'address': >> > {u'bus': u'0', u'controller': u'0', u'type': u'virtio-serial', u'port': >> > u'2'}}' found >> > Mar 24 19:57:58 bufferoverflow vdsm vm.Vm WARNING >> > vmId=`4d3d81b3-d083-4569-acc2-8e631ed51843`::_readPauseCode unsupported >> by >> > libvirt vm >> > Mar 24 19:57:58 bufferoverflow kernel: [ 7402.688177] ata1: hard >> resetting >> > link >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7402.994510] ata1: SATA link up >> > 6.0 Gbps (SStatus 133 SControl 300) >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.005510] ACPI Error: [DSSP] >> > Namespace lookup failure, AE_NOT_FOUND (20120711/psargs-359) >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.005517] ACPI Error: Method >> > parse/execution failed [\_SB_.PCI0.SAT0.SPT0._GTF] (Node >> 880407c74d48), >> > AE_NOT_FOUND (20120711/psparse-536) >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.015485] ACPI Error: [DSSP] >> > Namespace lookup failure, AE_NOT_FOUND (20120711/psargs-359) >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.015493] ACPI Error: Method >> > parse/execution failed [\_SB_.PCI0.SAT0.SPT0._GTF] (Node >> 880407c74d48), >> > AE_NOT_FOUND (20120711/psparse-536) >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.016061] ata1.00: >> configured >> > for UDMA/133 >> > Mar 24 19:57:59 bufferoverflow kernel: [ 7403.016066] ata1: EH complete >> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200 >> 7422 >> > [4759]: 1083422e close_task_aio 0 0x7ff3740008c0 busy >> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200 >> 7422 >> > [4759]: 1083422e close_task_aio 1 0x7ff374000910 busy >> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200 >> 7422 >> > [4759]: 1083422e close_task_aio 2 0x7ff374000960 busy >> > Mar 24 19:58:01 bufferoverflow sanlock[1208]: 2013-03-24 19:58:01+0200 >> 7422 >> > [4759]: 1083422e close_task_aio 3 0x7ff3740009b0 busy >> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.714145] device-mapper: >> table: >> > 253:0: multipath: error getting device >> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.714148] device-mapper: >> ioctl: >> > error adding target to table >> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.715051] device-mapper: >> table: >> > 253:0: multipath: error getting device >> > Mar 24 19:58:01 bufferoverflow kernel: [ 7405.715053] device-mapper: >> ioctl: >> > error
[Users] ovirt-iso-uploader
Hello, I am running ovirt 3.2.1 (built from sources) on Fedora 18. It seems that I don't have the ovirt-iso-uploader installed. [wil@bufferoverflow ovirt-engine]$ ovirt-iso-uploader list bash: ovirt-iso-uploader: command not found... [wil@bufferoverflow ovirt-engine]$ sudo yum list installed | grep -i uplo [wil@bufferoverflow ovirt-engine]$ What is the easiest way to install the ovirt-iso-uploader (preferably using "yum install")? Thanks, Limor G ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Failing to attach NFS data storage domain (Ovirt 3.2)
Updating from sanlock-2.6-4.fc18 to sanlock-2.6-7.fc18 fixed it. Huge thanks! On Mar 20, 2013 10:24 PM, "Federico Simoncelli" wrote: > - Original Message - > > From: "Limor Gavish" > > To: "Federico Simoncelli" > > Cc: "Yuval M" , users@ovirt.org, "Nezer Zaidenberg" < > nzaidenb...@mac.com>, "Eli Mesika" > > , "Maor Lipchuk" > > Sent: Wednesday, March 20, 2013 9:02:35 PM > > Subject: Re: [Users] Failing to attach NFS data storage domain (Ovirt > 3.2) > > > > Thank you very much for your response. > > > > > > Attached VDSM logs as you requested (The VDSM logs where the NFS > > domain was created were missing so we had to recreate the NFS > > domain, therefore the sdUUID has changed). > > Here is the rest of the commands you asked: > > > > > > [root@bufferoverflow wil]# mount -t nfs > > bufferoverflow:/home/BO_Ovirt_Storage /mnt/tmp > > [root@bufferoverflow wil]# cd > > /mnt/tmp/1083422e-a5db-41b6-b667-b9ef1ef244f0/dom_md/ > > [root@bufferoverflow dom_md]# ls -l > > total 2052 > > -rw-rw 1 vdsm kvm 1048576 Mar 20 21:46 ids > > -rw-rw 1 vdsm kvm 0 Mar 20 21:45 inbox > > -rw-rw 1 vdsm kvm 2097152 Mar 20 21:45 leases > > -rw-r--r-- 1 vdsm kvm 311 Mar 20 21:45 metadata > > -rw-rw 1 vdsm kvm 0 Mar 20 21:45 outbox > > [root@bufferoverflow dom_md]# sanlock direct dump ids > > Sorry I should have mentioned that if you use root_squash for your nfs > share you have to switch to the vdsm user: > > (root)# su -s /bin/sh vdsm > (vdsm)$ cd /mnt/tmp//dom_md/ > > (vdsm)$ sanlock direct dump ids > (and now you should be able to see the output) > > If the output is still empty then used hexdump -C to inspect it > (and eventually post it here compressed). > > Another important thing that you should check is: > > # ps fax | grep sanlock > > If the output doesn't look like the following: > > 1966 ?SLs0:00 wdmd -G sanlock > 2036 ?SLsl 0:00 sanlock daemon -U sanlock -G sanlock > 2037 ?S 0:00 \_ sanlock daemon -U sanlock -G sanlock > > Then I suggest you to update sanlock to the latest build: > > http://koji.fedoraproject.org/koji/buildinfo?buildID=377815 > (sanlock-2.6-7.fc18) > > And eventually if after rebooting the problem persists, please post > also the sanlock log (/var/log/sanlock.log) > > > Please note, the VDSM is running as a system service (it was > > installed from a package) while ovirt-engine was built from sources > > and thus is not running as root. Is this an issue? > > It shouldn't be. > > -- > Federico > ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users