Re: [ovirt-users] hosted engine health check issues
On 04/23/2014 12:28 AM, Doron Fediuck wrote: Hi Rene, any idea what closed your ovirtmgmt bridge? as long as it is down vdsm may have issues starting up properly and this is why you see the complaints on the rpc server. Can you try manually fixing the network part first and then restart vdsm? Once vdsm is happy hosted engine VM will start. Thanks for your feedback, Doron. My ovirtmgmt bridge seems to be on or isn't it: # brctl show ovirtmgmt bridge name bridge id STP enabled interfaces ovirtmgmt 8000.0025907587c2 no eth0.200 # ip a s ovirtmgmt 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UNKNOWN link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever # ip a s eth0.200 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UP link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever I tried the following yesterday: Copy virtual disk from GlusterFS storage to local disk of host and create a new vm with virt-manager which loads ovirtmgmt disk. I could reach my engine over the ovirtmgmt bridge (so bridge must be working). I also started libvirtd with Option -v and I saw the following in libvirtd.log when trying to start ovirt engine: 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : Command result 0, with PID 11491 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is not a chain So it could be that something is broken in my hosted-engine network. Do you have any clue how I can troubleshoot this? Thanks, René - Original Message - From: René Koch rk...@linuxland.at To: Martin Sivak msi...@redhat.com Cc: users@ovirt.org Sent: Tuesday, April 22, 2014 1:46:38 PM Subject: Re: [ovirt-users] hosted engine health check issues Hi, I rebooted one of my ovirt hosts today and the result is now that I can't start hosted-engine anymore. ovirt-ha-agent isn't running because the lockspace file is missing (sanlock complains about it). So I tried to start hosted-engine with --vm-start and I get the following errors: == /var/log/sanlock.log == 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 == /var/log/messages == Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering disabled state Apr 22 12:38:17 ovirt-host02 kernel: device vnet0 left promiscuous mode Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering disabled state == /var/log/vdsm/vdsm.log == Thread-21::DEBUG::2014-04-22 12:38:17,563::libvirtconnection::124::root::(wrapper) Unknown libvirterror: ecode: 38 edom: 42 level: 2 message: Failed to acquire lock: No space left on device Thread-21::DEBUG::2014-04-22 12:38:17,563::vm::2263::vm.Vm::(_startUnderlyingVm) vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::_ongoingCreations released Thread-21::ERROR::2014-04-22 12:38:17,564::vm::2289::vm.Vm::(_startUnderlyingVm) vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed Traceback (most recent call last): File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm self._run() File /usr/share/vdsm/vm.py, line 3170, in _run self._connection.createXML(domxml, flags), File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py, line 92, in wrapper ret = f(*args, **kwargs) File /usr/lib64/python2.6/site-packages/libvirt.py, line 2665, in createXML if ret is None:raise libvirtError('virDomainCreateXML() failed', conn=self) libvirtError: Failed to acquire lock: No space left on device == /var/log/messages == Apr 22 12:38:17 ovirt-host02 vdsm vm.Vm ERROR vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed#012Traceback (most recent call last):#012 File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm#012 self._run()#012 File /usr/share/vdsm/vm.py, line 3170, in _run#012 self._connection.createXML(domxml, flags),#012 File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py, line 92, in wrapper#012ret = f(*args, **kwargs)#012 File /usr/lib64/python2.6/site-packages/libvirt.py, line 2665, in createXML#012if ret is None:raise libvirtError('virDomainCreateXML() failed', conn=self)#012libvirtError: Failed to acquire lock: No space left on device == /var/log/vdsm/vdsm.log == Thread-21::DEBUG::2014-04-22 12:38:17,569::vm::2731::vm.Vm::(setDownStatus)
Re: [ovirt-users] is spice html5 console actually working
Did you import your engine CA (https://ovirt-engine.example.org/ca.crt) to firefox? David On Pá, 2014-04-18 at 16:07 -0500, Jeremiah Jahn wrote: I really seems to point to something strange that I've done with my certificates I guess.. I get the following on chrome WebSocket server settings: - Listen on *:6100 - Flash security policy server - SSL/TLS support - proxying from *:6100 to targets in /dummy 1: handler exception: WSRequestHandler instance has no attribute 'last_code' and the following on firefox: [root@bluejay X86]# - proxying from *:6100 to targets in /dummy 1: handler exception: WSRequestHandler instance has no attribute 'last_code' 2: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL routines:SSL3_READ_BYTES:sslv3 alert bad certificate 3: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL routines:SSL3_READ_BYTES:sslv3 alert bad certificate 4: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL routines:SSL3_READ_BYTES:tlsv1 alert access denied 5: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL routines:SSL3_READ_BYTES:tlsv1 alert access denied 6: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL routines:SSL3_READ_BYTES:tlsv1 alert access denied is there a way to just turn off SSL websockets for now? On Fri, Apr 18, 2014 at 9:24 AM, Jeremiah Jahn jerem...@goodinassociates.com wrote: Even after I've manually imported the ca.cert file into both chrome and firefox, I get the same error as before. Is there something that needs to be done on each host? I installed the websocket-proxy after the fact, and each host is a manually installed host, as opposed to the node.iso install. On Fri, Apr 18, 2014 at 8:28 AM, Jeremiah Jahn jerem...@goodinassociates.com wrote: And yet I always seem to get this error: WebSocket connection to 'wss://localhost:6100/eyJ2YWxpZFRvIjoiMjAxNDA0MTgxMzI4MjMiLCJkYXRhIjoiJTdCJT...lHbjZVRHhzOThaRndVdUlmM0s5Wk1VQVh4K2QwZz09Iiwic2FsdCI6IjlxZFkzakN4U0lNPSJ9' failed: WebSocket is closed before the connection is established. spicehtml5-main.html?host=localhostport=6100:1 disconnect spicehtml5-main.html?host=localhostport=6100:90 Error in event handler for (unknown): TypeError: Cannot read property 'state' of null at CSRecorder.onQueryStateCompleted (chrome-extension://cplklnmnlbnpmjogncfgfijoopmnlemp/content_scripts/recorder.js:43:13) at extensions::messaging:327:9 at Function.target.(anonymous function) (extensions::SafeBuiltins:19:14) at Event.dispatchToListener (extensions::event_bindings:386:22) at Event.dispatch_ (extensions::event_bindings:371:27) at Event.dispatch (extensions::event_bindings:392:17) at dispatchOnMessage (extensions::messaging:294:22) spicehtml5-main.html?host=localhostport=6100:1 On Thu, Apr 17, 2014 at 6:21 PM, Maurice James mja...@media-node.com wrote: There are a few steps. Download the CA cert from your manager https://ovirtaddress/ca.crt Make sure it is trusted. Make sure ovirt-webproxy-socket is installed and running. Sent from my Galaxy S(R)III Original message From: Jeremiah Jahn Date:04/17/2014 9:56 AM (GMT-05:00) To: users@ovirt.org Subject: [ovirt-users] is spice html5 console actually working Has anyone gotten the html5 spice console to work, and did you have to do anything special other than enable it? I've tried every browser except opera and ie on linux and mac ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Feature Page: Mac Pool per DC
Hi, I was describing current state, first iteration. Need of restart is something which should not exist, I've removed that necessity meantime. Altered flow: You allocate mac address for nic in data center without own pool, it gets registered in global pool. Then you modify settings of that data center so that new pool is created for it. All NICs for that data center is queries from DB, it's macs released from global pool and added to data center scope pool. And other way around. When you delete this scoped pool, all its content will be moved to global pool. Feature page is updated. Note: *previously* there was MAC placed in wrong pool only after modification of existing data center, which caused entirely new pool to be created (there wasn't pool for this scope, after modification there is). All other operations were fine. Now all manipulation with scoped pools should be ok. Note2: all that scoped pool handling is implemented as strategy. If we are unsatisfied with this implementation we could create another one and switch to it without modifying 'calling' code. Also many implementation may coexist and we can switch between them (on app start up) upon config. Question: When allocating MAC, not one specified by user, system picks available mac from given mac pool. Imagine, that after some time then mac pool ranges changes, and lets say that whole new interval of macs is used, not overlapping with former one. Then all previously allocated macs will be present in altered pool as a user specified ones -- since they are outside of defined ranges. With large number of this mac address this have detrimental effect on memory usage. So if this is a real scenario, it would be acceptable(or welcomed) for you to reassign all mac address which were selected by system? For example on engine start / vm start. M. - Original Message - From: Itamar Heim ih...@redhat.com To: Martin Mucha mmu...@redhat.com Cc: users@ovirt.org, de...@ovirt.org Sent: Tuesday, April 22, 2014 5:15:35 PM Subject: Re: [ovirt-users] Feature Page: Mac Pool per DC On 04/18/2014 01:17 PM, Martin Mucha wrote: Hi, I'll try to describe it little bit more. Lets say, that we've got one data center. It's not configured yet to have its own mac pool. So in system is only one, global pool. We create few VMs and it's NICs will obtain its MAC from this global pool, marking them as used. Next we alter data center definition, so now it uses it's own mac pool. In system from this point on exists two mac pools, one global and one related to this data center, but those allocated MACs are still allocated in global pool, since new data center creation does not (yet) contain logic to get all assigned MACs related to this data center and reassign them in new pool. However, after app restart all VmNics are read from db and placed to appropriate pools. Lets assume, that we've performed such restart. Now we realized, that we actually don't want that data center have own mac pool, so we alter it's definition removing mac pool ranges. Pool related to this data center will be removed and it's content will ! be moved t o a scope above this data center -- into global scope pool. We know, that everything what's allocated in pool to be removed is still used, but we need to track it elsewhere and currently there's just one option, global pool. So to answer your last question. When I remove scope, it's pool is gone and its content moved elsewhere. Next, when MAC is returned to the pool, the request goes like: give me pool for this virtual machine, and whatever pool it is, I'm returning this MAC to it. Clients of ScopedMacPoolManager do not know which pool they're talking to. Decision, which pool is right for them, is done behind the scenes upon their identification (I want pool for this logical network). Notice, that there is one problem in deciding which scope/pool to use. There are places in code, which requires pool related to given data center, identified by guid. For that request, only data center scope or something broader like global scope can be returned. So even if one want to use one pool per logical network, requests identified by data center id still can return only data center scope or broader, and there are no chance returning pool related to logical network (except for situation, where there is sole logical network in that data center). Thanks for suggestion for another scopes. One question: if we're implementing them, would you like just to pick a *sole* non-global scope you want to use in your system (like data center related pools ONLY plus one global, or logical network related pools ONLY plus one global) or would it be (more) beneficial to you to have implemented some sort of cascading and overriding? Like: this data center uses *this* pool, BUT except for *this* logical network, which should use *this* one instead. I'll update feature page to contain these
Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine
I think that you are mixing up a lot of different things and to be honest I am not sure what configuration you have and what exactly you are trying to do. so lets try to simplify it? what type of storage are you working on? which host is the spm? On 04/22/2014 07:36 PM, Tamer Lima wrote: hello, I am in trouble I have 3 servers dedicated to test OVIRT: 01- engine + vdsm (8 cpus, 32GB ram , 2TB HD) 02 - vdsm (8 cpus, 32GB ram , 2TB HD) 03 - vdsm (8 cpus, 32GB ram , 2TB HD) I want to create cloned virtual machines but in my configuration I can only save virtual machines on server 01; my configuration refers a DATA DOMAIN on server 01 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were created like CLONE My server 01 is the data domain and all new virtual machine is created, via NFS, on server 01 , who has 2TB maximum capacity ( the same size of partition /sda3 = 2TB) how can I save each virtual machine on a desired vdsm server ? What I want is : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 04 virtual machines running and hosted phisicaly on this host server 03 - vdsm : 04 virtual machines running and hosted phisicaly on this host but I have this : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 01 virtual machines running on this server BUT hosted phisicaly on server 01 server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB ) How to solve this problem ? is it possible create one DATA DOMAIN for each VDSM host ? I think this is the solution but I do not know how to point VMs to be saved on specific data domain. thanks On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote: hi, thanks for reply I am investigating what is and how thin virtualization works Do you know if HADOOP is indicated to work under thin environment ? On Hadoop I will put large workloads and this thin virtualization utilizes more resources than exists (shareable environment) that is, if I have a real physical necessity of 500gb for each hadoop host and my Thin Virtualization has 2TB on NFS, I can have only 4 virtual machines (500GB each), or less. For this case I believe clone virtual machine is the right choice. But in my environment it takes 1h30m to build one cloned virtual machine. if you plan to overcommit then go with thin. The drawback is that if you of course hit the physical limit the VMs will run out of space... if you plan to allocate 500GB each, consume all of it, never plan to grow then go with the clone….yes, it's going to take time to write all that stuff. With thin you need to do the same amount of writes, but gradually over time while you're allocating it it hope it helps Thanks, michal Am I correct ? On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote: Hello, I created VMs by two ways : 1) on tab virtual machines new vm template (centos_65_64bits) 1.1 configuration : I do not select stateless checkbox 1.2 this process takes a 1h30 to create each machine. 2) on tab pools new vm template (centos_65_64bits) 2.1 default configuration : stateless 2.2 Here I created 3 virtual machines at once 2.3 this process takes only one minute On the tab virtual machines I can see all virtual machines. Pooled machines have different icon image and description is different too: machines generated from tab VM are described as clone/dependent - clone is a phisical copy? machines generated from tab POOL are described as thin/independent - thin is a just a reference to template vm ? what is phisical? any configuration file? yeah, sort of. just google thin provisioning in general:) In practice, what is the difference between these machines ? http://www.ovirt.org/Features/PrestartedVm Today there are 2 types of Vm pools: • Manual - the Vm is supposed to be manually returned to the pool. In practice, this is not really entirely supported. • Automatic - once the user shuts down the Vm - it returns to the pool (stateless). all vm created from pool are stateless ? the automatic pool, yes Thanks, michal thanks
Re: [ovirt-users] hosted engine health check issues
Hi René, libvirtError: Failed to acquire lock: No space left on device 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Can you please check the contents of /rhev/data-center/your nfs mount/nfs domain uuid/ha_agent/? This is how it should look like: [root@dev-03 ~]# ls -al /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/ total 2036 drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 . drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 .. -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata The errors seem to indicate that you somehow lost the lockspace file. -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 12:28 AM, Doron Fediuck wrote: Hi Rene, any idea what closed your ovirtmgmt bridge? as long as it is down vdsm may have issues starting up properly and this is why you see the complaints on the rpc server. Can you try manually fixing the network part first and then restart vdsm? Once vdsm is happy hosted engine VM will start. Thanks for your feedback, Doron. My ovirtmgmt bridge seems to be on or isn't it: # brctl show ovirtmgmt bridge name bridge id STP enabled interfaces ovirtmgmt 8000.0025907587c2 no eth0.200 # ip a s ovirtmgmt 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UNKNOWN link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever # ip a s eth0.200 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UP link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever I tried the following yesterday: Copy virtual disk from GlusterFS storage to local disk of host and create a new vm with virt-manager which loads ovirtmgmt disk. I could reach my engine over the ovirtmgmt bridge (so bridge must be working). I also started libvirtd with Option -v and I saw the following in libvirtd.log when trying to start ovirt engine: 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : Command result 0, with PID 11491 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is not a chain So it could be that something is broken in my hosted-engine network. Do you have any clue how I can troubleshoot this? Thanks, René - Original Message - From: René Koch rk...@linuxland.at To: Martin Sivak msi...@redhat.com Cc: users@ovirt.org Sent: Tuesday, April 22, 2014 1:46:38 PM Subject: Re: [ovirt-users] hosted engine health check issues Hi, I rebooted one of my ovirt hosts today and the result is now that I can't start hosted-engine anymore. ovirt-ha-agent isn't running because the lockspace file is missing (sanlock complains about it). So I tried to start hosted-engine with --vm-start and I get the following errors: == /var/log/sanlock.log == 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 == /var/log/messages == Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering disabled state Apr 22 12:38:17 ovirt-host02 kernel: device vnet0 left promiscuous mode Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering disabled state == /var/log/vdsm/vdsm.log == Thread-21::DEBUG::2014-04-22 12:38:17,563::libvirtconnection::124::root::(wrapper) Unknown libvirterror: ecode: 38 edom: 42 level: 2 message: Failed to acquire lock: No space left on device Thread-21::DEBUG::2014-04-22 12:38:17,563::vm::2263::vm.Vm::(_startUnderlyingVm) vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::_ongoingCreations released Thread-21::ERROR::2014-04-22 12:38:17,564::vm::2289::vm.Vm::(_startUnderlyingVm) vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed Traceback (most recent call last): File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm self._run() File /usr/share/vdsm/vm.py, line 3170, in _run self._connection.createXML(domxml, flags), File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py, line 92, in wrapper ret = f(*args, **kwargs) File
Re: [ovirt-users] Feature Page: Mac Pool per DC
Hi, I'm not 100% sure I understand the question. What must be ensured: No VM, which got a manually assigned address, which is (not) in any configured pool, should stick to this address. I don't know if you plan to reassign some random address from the same (or any, if the mac is in no pool) pool. But this should be avoided, as you may rely on mac consistency for your vms. Am 23.04.2014 10:12, schrieb Martin Mucha: So if this is a real scenario, it would be acceptable(or welcomed) for you to reassign all mac address which were selected by system? -- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] hosted engine health check issues
On 04/23/2014 11:08 AM, Martin Sivak wrote: Hi René, libvirtError: Failed to acquire lock: No space left on device 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Can you please check the contents of /rhev/data-center/your nfs mount/nfs domain uuid/ha_agent/? This is how it should look like: [root@dev-03 ~]# ls -al /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/ total 2036 drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 . drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 .. -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata The errors seem to indicate that you somehow lost the lockspace file. True :) Isn't this file created when hosted engine is started? Or how can I create this file manually? -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 12:28 AM, Doron Fediuck wrote: Hi Rene, any idea what closed your ovirtmgmt bridge? as long as it is down vdsm may have issues starting up properly and this is why you see the complaints on the rpc server. Can you try manually fixing the network part first and then restart vdsm? Once vdsm is happy hosted engine VM will start. Thanks for your feedback, Doron. My ovirtmgmt bridge seems to be on or isn't it: # brctl show ovirtmgmt bridge name bridge id STP enabled interfaces ovirtmgmt 8000.0025907587c2 no eth0.200 # ip a s ovirtmgmt 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UNKNOWN link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever # ip a s eth0.200 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UP link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever I tried the following yesterday: Copy virtual disk from GlusterFS storage to local disk of host and create a new vm with virt-manager which loads ovirtmgmt disk. I could reach my engine over the ovirtmgmt bridge (so bridge must be working). I also started libvirtd with Option -v and I saw the following in libvirtd.log when trying to start ovirt engine: 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : Command result 0, with PID 11491 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is not a chain So it could be that something is broken in my hosted-engine network. Do you have any clue how I can troubleshoot this? Thanks, René - Original Message - From: René Koch rk...@linuxland.at To: Martin Sivak msi...@redhat.com Cc: users@ovirt.org Sent: Tuesday, April 22, 2014 1:46:38 PM Subject: Re: [ovirt-users] hosted engine health check issues Hi, I rebooted one of my ovirt hosts today and the result is now that I can't start hosted-engine anymore. ovirt-ha-agent isn't running because the lockspace file is missing (sanlock complains about it). So I tried to start hosted-engine with --vm-start and I get the following errors: == /var/log/sanlock.log == 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 == /var/log/messages == Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering disabled state Apr 22 12:38:17 ovirt-host02 kernel: device vnet0 left promiscuous mode Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering disabled state == /var/log/vdsm/vdsm.log == Thread-21::DEBUG::2014-04-22 12:38:17,563::libvirtconnection::124::root::(wrapper) Unknown libvirterror: ecode: 38 edom: 42 level: 2 message: Failed to acquire lock: No space left on device Thread-21::DEBUG::2014-04-22 12:38:17,563::vm::2263::vm.Vm::(_startUnderlyingVm) vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::_ongoingCreations released Thread-21::ERROR::2014-04-22 12:38:17,564::vm::2289::vm.Vm::(_startUnderlyingVm) vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed Traceback (most recent call last): File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm self._run() File /usr/share/vdsm/vm.py, line 3170, in _run self._connection.createXML(domxml, flags), File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py, line 92, in wrapper ret = f(*args, **kwargs) File
Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine
Hi, the solution for this would be either use gluster or another shared storage software across all 3 hosts or go with local storage datacenters. Am 22.04.2014 20:36, schrieb Tamer Lima: hello, I am in trouble I have 3 servers dedicated to test OVIRT: 01- engine + vdsm (8 cpus, 32GB ram , 2TB HD) 02 - vdsm (8 cpus, 32GB ram , 2TB HD) 03 - vdsm (8 cpus, 32GB ram , 2TB HD) I want to create cloned virtual machines but in my configuration I can only save virtual machines on server 01; my configuration refers a DATA DOMAIN on server 01 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were created like CLONE My server 01 is the data domain and all new virtual machine is created, via NFS, on server 01 , who has 2TB maximum capacity ( the same size of partition /sda3 = 2TB) how can I save each virtual machine on a desired vdsm server ? What I want is : server 01 - engine + vdsm :03 virtual machines running and hosted phisicaly on this host server 02 - vdsm :04 virtual machines running and hosted phisicaly on this host server 03 - vdsm :04 virtual machines running and hosted phisicaly on this host but I have this : server 01 - engine + vdsm :03 virtual machines running and hosted phisicaly on this host server 02 - vdsm :01 virtual machines running on this server BUT hosted phisicaly on server 01 server 03 - vdsm :none, because my DATA DOMAIN IS FULL (2TB ) How to solve this problem ? is it possible create one DATA DOMAIN for each VDSM host ? I think this is the solution but I do not know how to point VMs to be saved on specific data domain. thanks On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek michal.skriva...@redhat.com wrote: On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com wrote: hi, thanks for reply I am investigating what is and how thin virtualization works Do you know if HADOOP is indicated to work under thin environment ? On Hadoop I will put large workloads and this thin virtualization utilizes more resources than exists (shareable environment) that is, if I have a real physical necessity of 500gb for each hadoop host and my Thin Virtualization has 2TB on NFS, I can have only 4 virtual machines (500GB each), or less. For this case I believe clone virtual machine is the right choice. But in my environment it takes 1h30m to build one cloned virtual machine. if you plan to overcommit then go with thin. The drawback is that if you of course hit the physical limit the VMs will run out of space... if you plan to allocate 500GB each, consume all of it, never plan to grow then go with the clone….yes, it's going to take time to write all that stuff. With thin you need to do the same amount of writes, but gradually over time while you're allocating it it hope it helps Thanks, michal Am I correct ? On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek michal.skriva...@redhat.com wrote: On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com wrote: Hello, I created VMs by two ways : 1) on tab virtual machines new vm template (centos_65_64bits) 1.1 configuration : I do not select stateless checkbox 1.2 this process takes a 1h30 to create each machine. 2) on tab pools new vm template (centos_65_64bits) 2.1 default configuration : stateless 2.2 Here I created 3 virtual machines at once 2.3 this process takes only one minute On the tab virtual machines I can see all virtual machines. Pooled machines have different icon image and description is different too: machines generated from tab VM are described as clone/dependent - clone is a phisical copy? machines generated from tab POOL are described as thin/independent - thin is a just a reference to template vm ? what is phisical? any configuration file? yeah, sort of. just google thin provisioning in general:) In practice, what is the difference between these machines ? http://www.ovirt.org/Features/PrestartedVm Today there are 2 types of Vm pools: • Manual - the Vm is supposed to be manually returned to the pool. In practice, this is not really entirely supported. • Automatic - once the user shuts down the Vm - it returns to the pool (stateless). all vm created from pool are stateless ? the automatic pool, yes Thanks, michal thanks ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.:
Re: [ovirt-users] Ovirt snapshot failing on one VM
steve, I did not say that there is a limit. there is no limit and you can take a 1000 snapshots if you like, I simply said that I think that it would not be would a good practice to do so. I also did not say that this is your current problem with the vm so you are jumping to conclusions here. i simply explained how snapshots work which is that they are created in a chain, if there is a problem at a single point in time it would effect the rest of the snapshots below it. And that we query all images under the base Image so if you have a lot of them it would take a long time for the results to come back. as for your vm, since you fail to create a snapshot on only that vm it means that there is a problem in the current vm and it's chain. I can see when comparing the uuid's that the pool, domain, base image and last snapshots all exists in the rhev link. 2014-04-22 12:13:41,083 INFO [org.ovirt.engine.core.vdsbroker.irsbroker.CreateSnapshotVDSCommand] (pool-6-thread-49) [7ccaed5] -- createVolume parameters: sdUUID=95b9d922-4df7-4d3b-9bca-467e2fd9d573 spUUID=9497ef2c-8368-4c92-8d61-7f318a90748f imgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b size=21,474,836,480 bytes volFormat=COW volType=Sparse volUUID=0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936 descr= srcImgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b srcVolUUID=1a67de4b-aa1c-4436-baca-ca55726d54d7 lets see if it's possibly a cache issue - can you please restart vdsm on the hosts? On 04/22/2014 08:22 PM, Steve Dainard wrote: All snapshots are from before failure. That's a bit scary that there may be a 'too many snapshots' issue. I take snapshots for point in time consistency, and without the ability to collapse them while the vm is running I'm not sure what the best option is here. What is the recommended snapshot limit? Or maybe a better question; whats the intended use case for snapshots in ovirt? Export domain is currently unavailable, and without it active I can't disable it properly. # ls -tl /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b total 8517740 -rw-rw. 1 vdsm kvm97583104 Apr 22 14:03 1a67de4b-aa1c-4436-baca-ca55726d54d7 -rw-r--r--. 1 vdsm kvm 268 Apr 22 12:13 1a67de4b-aa1c-4436-baca-ca55726d54d7.meta -rw-r--r--. 1 vdsm kvm 272 Apr 22 01:06 87390b64-becd-4a6f-a4fc-d27655f59b64.meta -rw-rw. 1 vdsm kvm 1048576 Apr 22 01:04 1a67de4b-aa1c-4436-baca-ca55726d54d7.lease -rw-rw. 1 vdsm kvm 107413504 Apr 20 22:00 87390b64-becd-4a6f-a4fc-d27655f59b64 -rw-rw. 1 vdsm kvm 104267776 Apr 19 22:00 6f9fd451-6c82-4390-802c-9e23a7d89427 -rw-rw. 1 vdsm kvm 1048576 Apr 19 22:00 87390b64-becd-4a6f-a4fc-d27655f59b64.lease -rw-r--r--. 1 vdsm kvm 272 Apr 19 22:00 6f9fd451-6c82-4390-802c-9e23a7d89427.meta -rw-rw. 1 vdsm kvm 118358016 Apr 18 22:00 c298ce3b-ec6a-4526-9971-a769f4d3d69b -rw-rw. 1 vdsm kvm 1048576 Apr 18 22:00 6f9fd451-6c82-4390-802c-9e23a7d89427.lease -rw-r--r--. 1 vdsm kvm 272 Apr 18 22:00 c298ce3b-ec6a-4526-9971-a769f4d3d69b.meta -rw-rw. 1 vdsm kvm 120913920 Apr 17 22:00 0ee58208-6be8-4f81-bd51-0bd4b6d5d83a -rw-rw. 1 vdsm kvm 1048576 Apr 17 22:00 c298ce3b-ec6a-4526-9971-a769f4d3d69b.lease -rw-r--r--. 1 vdsm kvm 272 Apr 17 22:00 0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.meta -rw-rw. 1 vdsm kvm 117374976 Apr 16 22:00 9aeb973d-9a54-441e-9ce9-f4f1a233da26 -rw-rw. 1 vdsm kvm 1048576 Apr 16 22:00 0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.lease -rw-r--r--. 1 vdsm kvm 272 Apr 16 22:00 9aeb973d-9a54-441e-9ce9-f4f1a233da26.meta -rw-rw. 1 vdsm kvm 110886912 Apr 15 22:00 0eae2185-884a-44d3-9099-e952b6b7ec37 -rw-rw. 1 vdsm kvm 1048576 Apr 15 22:00 9aeb973d-9a54-441e-9ce9-f4f1a233da26.lease -rw-r--r--. 1 vdsm kvm 272 Apr 15 22:00 0eae2185-884a-44d3-9099-e952b6b7ec37.meta -rw-rw. 1 vdsm kvm 1048576 Apr 14 22:00 0eae2185-884a-44d3-9099-e952b6b7ec37.lease -rw-rw. 1 vdsm kvm 164560896 Apr 14 22:00 ceffc643-b823-44b3-961e-93f3dc971886 -rw-r--r--. 1 vdsm kvm 272 Apr 14 22:00 ceffc643-b823-44b3-961e-93f3dc971886.meta -rw-rw. 1 vdsm kvm 1048576 Apr 13 22:00 ceffc643-b823-44b3-961e-93f3dc971886.lease -rw-r--r--. 1 vdsm kvm 272 Apr 13 22:00 878fc690-ab08-489c-955b-9159f62026b1.meta -rw-rw. 1 vdsm kvm 109182976 Apr 13 21:59 878fc690-ab08-489c-955b-9159f62026b1 -rw-rw. 1 vdsm kvm 110297088 Apr 12 22:00 5210eec2-a0eb-462e-95d5-7cf27db312f5 -rw-rw. 1 vdsm kvm 1048576 Apr 12 22:00 878fc690-ab08-489c-955b-9159f62026b1.lease -rw-r--r--. 1 vdsm kvm 272 Apr 12 22:00 5210eec2-a0eb-462e-95d5-7cf27db312f5.meta -rw-rw. 1 vdsm kvm76480512 Apr 11 22:00
Re: [ovirt-users] Feature Page: Mac Pool per DC
Sorry, I meant every vm not No VM. Am 23.04.2014 11:08, schrieb Sven Kieske: No VM, which got a manually assigned address, which is (not) in any configured pool, should stick to this address. -- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Hosted Engine error -243
Hi, /var/log/ovirt-hosted-engine-ha/broker.log Host1: Thread-118327::INFO::2014-04-23 12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-118327::INFO::2014-04-23 12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle) Connection closed Thread-118328::INFO::2014-04-23 12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-118328::INFO::2014-04-23 12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle) Connection closed Host2: Thread-4::INFO::2014-04-23 12:36:08,020::mem_free::53::mem_free.MemFree::(action) memFree: 9816 Thread-3::INFO::2014-04-23 12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge ::(action) Found bridge ovirtmgmt Thread-296455::INFO::2014-04-23 12:36:08,678::listener::134::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-296455::INFO::2014-04-23 12:36:08,684::listener::184::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(handle) Connection closed /var/log/ovirt-hosted-engine-ha/agent.log host1: MainThread::INFO::2014-04-02 17:46:14,856::state_decorators::25::ovirt_hosted_en gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local engine vm statusno actions taken MainThread::INFO::2014-04-02 17:46:14,857::brokerlink::108::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1396453574.86 type=st ate_transition detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o virt.lan' MainThread::INFO::2014-04-02 17:46:14,858::brokerlink::117::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of state_transi tion (UnknownLocalVmState-UnknownLocalVmState) sent? ignored MainThread::WARNING::2014-04-02 17:46:15,463::hosted_engine::334::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error while monito ring engine: float() argument must be a string or a number MainThread::WARNING::2014-04-02 17:46:15,464::hosted_engine::337::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Unexpected error Traceback (most recent call last): File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/hosted_eng ine.py, line 323, in start_monitoring state.score(self._log)) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 160, in score lm, logger, score, score_cfg) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 61, in _penalize_memory if self._float_or_default(lm['mem-free'], 0) vm_mem: File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 51, in _float_or_default return float(value) TypeError: float() argument must be a string or a number MainThread::ERROR::2014-04-02 17:46:15,464::hosted_engine::350::ovirt_hosted_eng ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Shutting down the ag ent because of 3 failures in a row! MainThread::INFO::2014-04-02 17:46:15,466::agent::116:: ovirt_hosted_engine_ha.ag ent.agent.Agent::(run) Agent shutting down host2: MainThread::INFO::2014-04-23 12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Current state EngineUnexpectedlyDown (score: 0) MainThread::INFO::2014-04-23 12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1398249414.84 type=state_transition detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown hostname='host02.ovirt.lan' MainThread::INFO::2014-04-23 12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of state_transition (EngineUnexpectedlyDown-EngineUnexpectedlyDown) sent? ignored /var/log/vdsm/vdsm.log host1 : Thread-116::DEBUG::2014-04-23 12:40:17,060::fileSD::225::Storage.Misc.excCmd::(getReadDelay) '/bin/dd iflag=direct if=/rhev/data-center/mnt/host01.ovirt.lan:_home_iso/cc51143e-8ad7-4b0b-a4d2-9024dffc1188/dom_md/metadata bs=4096 count=1' (cwd None) Thread-116::DEBUG::2014-04-23 12:40:17,070::fileSD::225::Storage.Misc.excCmd::(getReadDelay) SUCCESS: err = '0+1 records in\n0+1 records out\n343 bytes (343 B) copied, 0.000183642 s, 1.9 MB/s\n'; rc = 0 Thread-37::DEBUG::2014-04-23 12:40:17,504::fileSD::225::Storage.Misc.excCmd::(getReadDelay) '/bin/dd iflag=direct if=/rhev/data-center/mnt/host01.ovirt.lan:_home_NFS01/aea040f8-ab9d-435b-9ecf-ddd4272e592f/dom_md/metadata bs=4096 count=1' (cwd None) Thread-37::DEBUG::2014-04-23
Re: [ovirt-users] Hosted Engine error -243
Hi, I'm not sure yet what causes the problem, but the workaround should be: open file /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py in your favorite editor, go to line 52 and change it: from: except ValueError: to: except (ValueError, TypeError): --Jirka On 04/23/2014 12:43 PM, Kevin Tibi wrote: Hi, /var/log/ovirt-hosted-engine-ha/broker.log Host1: Thread-118327::INFO::2014-04-23 12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-118327::INFO::2014-04-23 12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle) Connection closed Thread-118328::INFO::2014-04-23 12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-118328::INFO::2014-04-23 12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle) Connection closed Host2: Thread-4::INFO::2014-04-23 12:36:08,020::mem_free::53::mem_free.MemFree::(action ) memFree: 9816 Thread-3::INFO::2014-04-23 12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge ::(action) Found bridge ovirtmgmt Thread-296455::INFO::2014-04-23 12:36:08,678::listener::134::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-296455::INFO::2014-04-23 12:36:08,684::listener::184::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(handle) Connection closed /var/log/ovirt-hosted-engine-ha/agent.log host1: MainThread::INFO::2014-04-02 17:46:14,856::state_decorators::25::ovirt_hosted_en gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local engine vm statusno actions taken MainThread::INFO::2014-04-02 17:46:14,857::brokerlink::108::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1396453574.86 type=st ate_transition detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o virt.lan' MainThread::INFO::2014-04-02 17:46:14,858::brokerlink::117::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of state_transi tion (UnknownLocalVmState-UnknownLocalVmState) sent? ignored MainThread::WARNING::2014-04-02 17:46:15,463::hosted_engine::334::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error while monito ring engine: float() argument must be a string or a number MainThread::WARNING::2014-04-02 17:46:15,464::hosted_engine::337::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Unexpected error Traceback (most recent call last): File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/hosted_eng ine.py, line 323, in start_monitoring state.score(self._log)) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 160, in score lm, logger, score, score_cfg) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 61, in _penalize_memory if self._float_or_default(lm['mem-free'], 0) vm_mem: File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 51, in _float_or_default return float(value) TypeError: float() argument must be a string or a number MainThread::ERROR::2014-04-02 17:46:15,464::hosted_engine::350::ovirt_hosted_eng ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Shutting down the ag ent because of 3 failures in a row! MainThread::INFO::2014-04-02 17:46:15,466::agent::116::ovirt_hosted_engine_ha.ag http://ovirt_hosted_engine_ha.ag ent.agent.Agent::(run) Agent shutting down host2: MainThread::INFO::2014-04-23 12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Current state EngineUnexpectedlyDown (score: 0) MainThread::INFO::2014-04-23 12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1398249414.84 type=state_transition detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown hostname='host02.ovirt.lan' MainThread::INFO::2014-04-23 12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of state_transition (EngineUnexpectedlyDown-EngineUnexpectedlyDown) sent? ignored /var/log/vdsm/vdsm.log host1 : Thread-116::DEBUG::2014-04-23 12:40:17,060::fileSD::225::Storage.Misc.excCmd::(getReadDelay) '/bin/dd iflag=direct if=/rhev/data-center/mnt/host01.ovirt.lan:_home_iso/cc51143e-8ad7-4b0b-a4d2-9024dffc1188/dom_md/metadata bs=4096 count=1' (cwd None) Thread-116::DEBUG::2014-04-23 12:40:17,070::fileSD::225::Storage.Misc.excCmd::(getReadDelay) SUCCESS: err = '0+1
Re: [ovirt-users] hosted engine health check issues
Hi, Isn't this file created when hosted engine is started? The file is created by the setup script. If it got lost then there was probably something bad happening in your NFS or Gluster storage. Or how can I create this file manually? I can give you experimental treatment for this. We do not have any official way as this is something that should not ever happen :) !! But before you do that make sure you do not have any nodes running properly. This will destroy and reinitialize the lockspace database for the whole hosted-engine environment (which you apparently lack, but..). !! You have to create the ha_agent/hosted-engine.lockspace file with the expected size (1MB) and then tell sanlock to initialize it as a lockspace using: # python import sanlock sanlock.write_lockspace(lockspace=hosted-engine, ... path=/rhev/data-center/mnt/nfs/hosted engine storage domain/ha_agent/hosted-engine.lockspace, ... offset=0) Then try starting the services (both broker and agent) again. -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 11:08 AM, Martin Sivak wrote: Hi René, libvirtError: Failed to acquire lock: No space left on device 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Can you please check the contents of /rhev/data-center/your nfs mount/nfs domain uuid/ha_agent/? This is how it should look like: [root@dev-03 ~]# ls -al /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/ total 2036 drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 . drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 .. -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata The errors seem to indicate that you somehow lost the lockspace file. True :) Isn't this file created when hosted engine is started? Or how can I create this file manually? -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 12:28 AM, Doron Fediuck wrote: Hi Rene, any idea what closed your ovirtmgmt bridge? as long as it is down vdsm may have issues starting up properly and this is why you see the complaints on the rpc server. Can you try manually fixing the network part first and then restart vdsm? Once vdsm is happy hosted engine VM will start. Thanks for your feedback, Doron. My ovirtmgmt bridge seems to be on or isn't it: # brctl show ovirtmgmt bridge namebridge id STP enabled interfaces ovirtmgmt 8000.0025907587c2 no eth0.200 # ip a s ovirtmgmt 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UNKNOWN link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever # ip a s eth0.200 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UP link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever I tried the following yesterday: Copy virtual disk from GlusterFS storage to local disk of host and create a new vm with virt-manager which loads ovirtmgmt disk. I could reach my engine over the ovirtmgmt bridge (so bridge must be working). I also started libvirtd with Option -v and I saw the following in libvirtd.log when trying to start ovirt engine: 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : Command result 0, with PID 11491 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is not a chain So it could be that something is broken in my hosted-engine network. Do you have any clue how I can troubleshoot this? Thanks, René - Original Message - From: René Koch rk...@linuxland.at To: Martin Sivak msi...@redhat.com Cc: users@ovirt.org Sent: Tuesday, April 22, 2014 1:46:38 PM Subject: Re: [ovirt-users] hosted engine health check issues Hi, I rebooted one of my ovirt hosts today and the result is now that I can't start hosted-engine anymore. ovirt-ha-agent isn't running because the lockspace file is missing (sanlock complains about it). So I tried to start hosted-engine with --vm-start and I get the following errors: == /var/log/sanlock.log == 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 == /var/log/messages == Apr 22 12:38:17 ovirt-host02 sanlock[3079]:
Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine
hello, thanks for reply my storage is NFS v3, defined on host 01. My DATA-DOMAIN and ISO-DOMAIN are hosted on host 01; my SPM is located on host 03, I dont remember why. I tried to migrate SPM to host 01 but is not possible. All creation of virtual machine starts on server 01 ( On Wed, Apr 23, 2014 at 5:19 AM, Dafna Ron d...@redhat.com wrote: I think that you are mixing up a lot of different things and to be honest I am not sure what configuration you have and what exactly you are trying to do. so lets try to simplify it? what type of storage are you working on? which host is the spm? On 04/22/2014 07:36 PM, Tamer Lima wrote: hello, I am in trouble I have 3 servers dedicated to test OVIRT: 01- engine + vdsm (8 cpus, 32GB ram , 2TB HD) 02 - vdsm (8 cpus, 32GB ram , 2TB HD) 03 - vdsm (8 cpus, 32GB ram , 2TB HD) I want to create cloned virtual machines but in my configuration I can only save virtual machines on server 01; my configuration refers a DATA DOMAIN on server 01 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were created like CLONE My server 01 is the data domain and all new virtual machine is created, via NFS, on server 01 , who has 2TB maximum capacity ( the same size of partition /sda3 = 2TB) how can I save each virtual machine on a desired vdsm server ? What I want is : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 04 virtual machines running and hosted phisicaly on this host server 03 - vdsm : 04 virtual machines running and hosted phisicaly on this host but I have this : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 01 virtual machines running on this server BUT hosted phisicaly on server 01 server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB ) How to solve this problem ? is it possible create one DATA DOMAIN for each VDSM host ? I think this is the solution but I do not know how to point VMs to be saved on specific data domain. thanks On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote: hi, thanks for reply I am investigating what is and how thin virtualization works Do you know if HADOOP is indicated to work under thin environment ? On Hadoop I will put large workloads and this thin virtualization utilizes more resources than exists (shareable environment) that is, if I have a real physical necessity of 500gb for each hadoop host and my Thin Virtualization has 2TB on NFS, I can have only 4 virtual machines (500GB each), or less. For this case I believe clone virtual machine is the right choice. But in my environment it takes 1h30m to build one cloned virtual machine. if you plan to overcommit then go with thin. The drawback is that if you of course hit the physical limit the VMs will run out of space... if you plan to allocate 500GB each, consume all of it, never plan to grow then go with the clone….yes, it's going to take time to write all that stuff. With thin you need to do the same amount of writes, but gradually over time while you're allocating it it hope it helps Thanks, michal Am I correct ? On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote: Hello, I created VMs by two ways : 1) on tab virtual machines new vm template (centos_65_64bits) 1.1 configuration : I do not select stateless checkbox 1.2 this process takes a 1h30 to create each machine. 2) on tab pools new vm template (centos_65_64bits) 2.1 default configuration : stateless 2.2 Here I created 3 virtual machines at once 2.3 this process takes only one minute On the tab virtual machines I can see all virtual machines. Pooled machines have different icon image and description is different too: machines generated from tab VM are described as clone/dependent - clone is a phisical copy? machines generated from tab POOL are described as thin/independent - thin is a just a reference to template vm ? what is phisical? any configuration file? yeah, sort of. just google thin provisioning in general:) In practice, what is the difference between these machines ?
Re: [ovirt-users] is spice html5 console actually working
yes, yes I did... It's as if installing the webproxy after the fact caused it to self sign with it's own certificate or something. I edited the two html files so that for the moment it's not using https and that works fine, but I just can't get the https to work at all. On Wed, Apr 23, 2014 at 2:51 AM, David Jaša dj...@redhat.com wrote: Did you import your engine CA (https://ovirt-engine.example.org/ca.crt) to firefox? David On Pá, 2014-04-18 at 16:07 -0500, Jeremiah Jahn wrote: I really seems to point to something strange that I've done with my certificates I guess.. I get the following on chrome WebSocket server settings: - Listen on *:6100 - Flash security policy server - SSL/TLS support - proxying from *:6100 to targets in /dummy 1: handler exception: WSRequestHandler instance has no attribute 'last_code' and the following on firefox: [root@bluejay X86]# - proxying from *:6100 to targets in /dummy 1: handler exception: WSRequestHandler instance has no attribute 'last_code' 2: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL routines:SSL3_READ_BYTES:sslv3 alert bad certificate 3: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL routines:SSL3_READ_BYTES:sslv3 alert bad certificate 4: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL routines:SSL3_READ_BYTES:tlsv1 alert access denied 5: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL routines:SSL3_READ_BYTES:tlsv1 alert access denied 6: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL routines:SSL3_READ_BYTES:tlsv1 alert access denied is there a way to just turn off SSL websockets for now? On Fri, Apr 18, 2014 at 9:24 AM, Jeremiah Jahn jerem...@goodinassociates.com wrote: Even after I've manually imported the ca.cert file into both chrome and firefox, I get the same error as before. Is there something that needs to be done on each host? I installed the websocket-proxy after the fact, and each host is a manually installed host, as opposed to the node.iso install. On Fri, Apr 18, 2014 at 8:28 AM, Jeremiah Jahn jerem...@goodinassociates.com wrote: And yet I always seem to get this error: WebSocket connection to 'wss://localhost:6100/eyJ2YWxpZFRvIjoiMjAxNDA0MTgxMzI4MjMiLCJkYXRhIjoiJTdCJT...lHbjZVRHhzOThaRndVdUlmM0s5Wk1VQVh4K2QwZz09Iiwic2FsdCI6IjlxZFkzakN4U0lNPSJ9' failed: WebSocket is closed before the connection is established. spicehtml5-main.html?host=localhostport=6100:1 disconnect spicehtml5-main.html?host=localhostport=6100:90 Error in event handler for (unknown): TypeError: Cannot read property 'state' of null at CSRecorder.onQueryStateCompleted (chrome-extension://cplklnmnlbnpmjogncfgfijoopmnlemp/content_scripts/recorder.js:43:13) at extensions::messaging:327:9 at Function.target.(anonymous function) (extensions::SafeBuiltins:19:14) at Event.dispatchToListener (extensions::event_bindings:386:22) at Event.dispatch_ (extensions::event_bindings:371:27) at Event.dispatch (extensions::event_bindings:392:17) at dispatchOnMessage (extensions::messaging:294:22) spicehtml5-main.html?host=localhostport=6100:1 On Thu, Apr 17, 2014 at 6:21 PM, Maurice James mja...@media-node.com wrote: There are a few steps. Download the CA cert from your manager https://ovirtaddress/ca.crt Make sure it is trusted. Make sure ovirt-webproxy-socket is installed and running. Sent from my Galaxy S(R)III Original message From: Jeremiah Jahn Date:04/17/2014 9:56 AM (GMT-05:00) To: users@ovirt.org Subject: [ovirt-users] is spice html5 console actually working Has anyone gotten the html5 spice console to work, and did you have to do anything special other than enable it? I've tried every browser except opera and ie on linux and mac ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine
what do you mean that host1 is engine + vdsm, are you using hosted engine? On 04/23/2014 01:59 PM, Tamer Lima wrote: hello, thanks for reply my storage is NFS v3, defined on host 01. My DATA-DOMAIN and ISO-DOMAIN are hosted on host 01; my SPM is located on host 03, I dont remember why. I tried to migrate SPM to host 01 but is not possible. All creation of virtual machine starts on server 01 ( On Wed, Apr 23, 2014 at 5:19 AM, Dafna Ron d...@redhat.com mailto:d...@redhat.com wrote: I think that you are mixing up a lot of different things and to be honest I am not sure what configuration you have and what exactly you are trying to do. so lets try to simplify it? what type of storage are you working on? which host is the spm? On 04/22/2014 07:36 PM, Tamer Lima wrote: hello, I am in trouble I have 3 servers dedicated to test OVIRT: 01- engine + vdsm (8 cpus, 32GB ram , 2TB HD) 02 - vdsm (8 cpus, 32GB ram , 2TB HD) 03 - vdsm (8 cpus, 32GB ram , 2TB HD) I want to create cloned virtual machines but in my configuration I can only save virtual machines on server 01; my configuration refers a DATA DOMAIN on server 01 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were created like CLONE My server 01 is the data domain and all new virtual machine is created, via NFS, on server 01 , who has 2TB maximum capacity ( the same size of partition /sda3 = 2TB) how can I save each virtual machine on a desired vdsm server ? What I want is : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 04 virtual machines running and hosted phisicaly on this host server 03 - vdsm : 04 virtual machines running and hosted phisicaly on this host but I have this : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 01 virtual machines running on this server BUT hosted phisicaly on server 01 server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB ) How to solve this problem ? is it possible create one DATA DOMAIN for each VDSM host ? I think this is the solution but I do not know how to point VMs to be saved on specific data domain. thanks On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote: hi, thanks for reply I am investigating what is and how thin virtualization works Do you know if HADOOP is indicated to work under thin environment ? On Hadoop I will put large workloads and this thin virtualization utilizes more resources than exists (shareable environment) that is, if I have a real physical necessity of 500gb for each hadoop host and my Thin Virtualization has 2TB on NFS, I can have only 4 virtual machines (500GB each), or less. For this case I believe clone virtual machine is the right choice. But in my environment it takes 1h30m to build one cloned virtual machine. if you plan to overcommit then go with thin. The drawback is that if you of course hit the physical limit the VMs will run out of space... if you plan to allocate 500GB each, consume all of it, never plan to grow then go with the clone….yes, it's going to take time to write all that stuff. With thin you need to do the same amount of writes, but gradually over time while you're allocating it it hope it helps Thanks, michal Am I correct ? On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote:
[ovirt-users] Add a Direct Lun via rest API (Oivrt 3.3.5)
Hello I was just wondering if anyone would be able to help me figure out if there is a way to login to an ISCSI target (EqualLogic) and add its associated volume as a Direct LUN via the REST api. I have figured out how to add an existing Direct LUN to a vm. I have created a volume on the SAN and then I am attempting to upload some xml to the API: curl -v -u 'admin@internal:mypass' -H Content-type: application/xml -d @disk.xml https://ovirt-test/disks/ --insecure cat disk.xml disk aliasdirect_lun/alias interfacevirtio/interface formatraw/format lunStorage typeiscsi/type logical_unit address10.0.0.1/address port3260/port targetiqn.2001-05.com.equallogic:0-1cb196-cff1c713e-e2a004dfcc65357b-dev-directlun/target /logical_unit /lunStorage /disk At the moment the API is returning with a HTTP 400: fault reasonIncomplete parameters/reason detailDisk [provisionedSize|size] required for add/detail /fault Is it possible to achieve my goal via the API ? Thanks *Gary Lloyd* -- IT Services Keele University --- ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine
hi, no. is not hosted engine. on server 01 I installed ovirt engine , vdsm, libvirt, postgres , etc using engine-setup . In this machine I have 3 VMs, all running centos65 this server 01 has ovirt web admin (jboss) and there I see all hosts ( server 01, 02 and 03) on the cluster. I can create and manage VM (thin and cloned) tamer Em 23/04/2014, às 10:21, Dafna Ron d...@redhat.com escreveu: what do you mean that host1 is engine + vdsm, are you using hosted engine? On 04/23/2014 01:59 PM, Tamer Lima wrote: hello, thanks for reply my storage is NFS v3, defined on host 01. My DATA-DOMAIN and ISO-DOMAIN are hosted on host 01; my SPM is located on host 03, I dont remember why. I tried to migrate SPM to host 01 but is not possible. All creation of virtual machine starts on server 01 ( On Wed, Apr 23, 2014 at 5:19 AM, Dafna Ron d...@redhat.com mailto:d...@redhat.com wrote: I think that you are mixing up a lot of different things and to be honest I am not sure what configuration you have and what exactly you are trying to do. so lets try to simplify it? what type of storage are you working on? which host is the spm? On 04/22/2014 07:36 PM, Tamer Lima wrote: hello, I am in trouble I have 3 servers dedicated to test OVIRT: 01- engine + vdsm (8 cpus, 32GB ram , 2TB HD) 02 - vdsm (8 cpus, 32GB ram , 2TB HD) 03 - vdsm (8 cpus, 32GB ram , 2TB HD) I want to create cloned virtual machines but in my configuration I can only save virtual machines on server 01; my configuration refers a DATA DOMAIN on server 01 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were created like CLONE My server 01 is the data domain and all new virtual machine is created, via NFS, on server 01 , who has 2TB maximum capacity ( the same size of partition /sda3 = 2TB) how can I save each virtual machine on a desired vdsm server ? What I want is : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 04 virtual machines running and hosted phisicaly on this host server 03 - vdsm : 04 virtual machines running and hosted phisicaly on this host but I have this : server 01 - engine + vdsm : 03 virtual machines running and hosted phisicaly on this host server 02 - vdsm : 01 virtual machines running on this server BUT hosted phisicaly on server 01 server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB ) How to solve this problem ? is it possible create one DATA DOMAIN for each VDSM host ? I think this is the solution but I do not know how to point VMs to be saved on specific data domain. thanks On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote: On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com wrote: hi, thanks for reply I am investigating what is and how thin virtualization works Do you know if HADOOP is indicated to work under thin environment ? On Hadoop I will put large workloads and this thin virtualization utilizes more resources than exists (shareable environment) that is, if I have a real physical necessity of 500gb for each hadoop host and my Thin Virtualization has 2TB on NFS, I can have only 4 virtual machines (500GB each), or less. For this case I believe clone virtual machine is the right choice. But in my environment it takes 1h30m to build one cloned virtual machine. if you plan to overcommit then go with thin. The drawback is that if you of course hit the physical limit the VMs will run out of space... if you plan to allocate 500GB each, consume all of it, never plan to grow then go with the clone….yes, it's going to take time to write all that stuff. With thin you need to do the same amount of writes, but gradually over time while you're allocating it it hope it helps Thanks, michal Am I correct ? On Thu, Apr 17, 2014 at 7:33 AM,
[ovirt-users] Unable to move disk to another data domain
Hi All, I am trying to move the disk of VM from Data Domain(Master) to another data domain in the same cluster, but I get the warning as some of the disk cannot be moved and I cannot see the list of other data domain in the selection window. Can someone help here ? Regards, Yusuf ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Hosted Engine error -243
In engine, i have Hosted Engine HA: not activefor my host1 Hosted Engine HA: active (score 0) for my host2 2014-04-23 13:52 GMT+02:00 Jiri Moskovcak jmosk...@redhat.com: Hi, I'm not sure yet what causes the problem, but the workaround should be: open file /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py in your favorite editor, go to line 52 and change it: from: except ValueError: to: except (ValueError, TypeError): --Jirka On 04/23/2014 12:43 PM, Kevin Tibi wrote: Hi, /var/log/ovirt-hosted-engine-ha/broker.log Host1: Thread-118327::INFO::2014-04-23 12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(setup) Connection established Thread-118327::INFO::2014-04-23 12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(handle) Connection closed Thread-118328::INFO::2014-04-23 12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(setup) Connection established Thread-118328::INFO::2014-04-23 12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(handle) Connection closed Host2: Thread-4::INFO::2014-04-23 12:36:08,020::mem_free::53::mem_free.MemFree::(action ) memFree: 9816 Thread-3::INFO::2014-04-23 12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge ::(action) Found bridge ovirtmgmt Thread-296455::INFO::2014-04-23 12:36:08,678::listener::134::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-296455::INFO::2014-04-23 12:36:08,684::listener::184::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(handle) Connection closed /var/log/ovirt-hosted-engine-ha/agent.log host1: MainThread::INFO::2014-04-02 17:46:14,856::state_decorators::25::ovirt_hosted_en gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local engine vm statusno actions taken MainThread::INFO::2014-04-02 17:46:14,857::brokerlink::108::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1396453574.86 type=st ate_transition detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o virt.lan' MainThread::INFO::2014-04-02 17:46:14,858::brokerlink::117::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of state_transi tion (UnknownLocalVmState-UnknownLocalVmState) sent? ignored MainThread::WARNING::2014-04-02 17:46:15,463::hosted_engine::334::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error while monito ring engine: float() argument must be a string or a number MainThread::WARNING::2014-04-02 17:46:15,464::hosted_engine::337::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Unexpected error Traceback (most recent call last): File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/hosted_eng ine.py, line 323, in start_monitoring state.score(self._log)) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 160, in score lm, logger, score, score_cfg) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 61, in _penalize_memory if self._float_or_default(lm['mem-free'], 0) vm_mem: File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py , line 51, in _float_or_default return float(value) TypeError: float() argument must be a string or a number MainThread::ERROR::2014-04-02 17:46:15,464::hosted_engine::350::ovirt_hosted_eng ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Shutting down the ag ent because of 3 failures in a row! MainThread::INFO::2014-04-02 17:46:15,466::agent::116::ovirt_hosted_engine_ha.ag http://ovirt_hosted_engine_ha.ag ent.agent.Agent::(run) Agent shutting down host2: MainThread::INFO::2014-04-23 12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha. agent.hosted_engine.HostedEngine::(start_monitoring) Current state EngineUnexpectedlyDown (score: 0) MainThread::INFO::2014-04-23 12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib. brokerlink.BrokerLink::(notify) Trying: notify time=1398249414.84 type=state_transition detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown hostname='host02.ovirt.lan' MainThread::INFO::2014-04-23 12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib. brokerlink.BrokerLink::(notify) Success, was notification of state_transition (EngineUnexpectedlyDown-EngineUnexpectedlyDown) sent? ignored /var/log/vdsm/vdsm.log host1 : Thread-116::DEBUG::2014-04-23
Re: [ovirt-users] Hosted Engine error -243
Ho god, My cpu usage is 80% on host1 1729 vdsm 20 0 762m 15m 2884 S 297.6 0.1 77:16.70 ovirt-ha-broker 2014-04-23 16:40 GMT+02:00 Kevin Tibi kevint...@hotmail.com: In engine, i have Hosted Engine HA: not activefor my host1 Hosted Engine HA: active (score 0) for my host2 2014-04-23 13:52 GMT+02:00 Jiri Moskovcak jmosk...@redhat.com: Hi, I'm not sure yet what causes the problem, but the workaround should be: open file /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py in your favorite editor, go to line 52 and change it: from: except ValueError: to: except (ValueError, TypeError): --Jirka On 04/23/2014 12:43 PM, Kevin Tibi wrote: Hi, /var/log/ovirt-hosted-engine-ha/broker.log Host1: Thread-118327::INFO::2014-04-23 12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(setup) Connection established Thread-118327::INFO::2014-04-23 12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(handle) Connection closed Thread-118328::INFO::2014-04-23 12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(setup) Connection established Thread-118328::INFO::2014-04-23 12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker. listener.ConnectionHandler::(handle) Connection closed Host2: Thread-4::INFO::2014-04-23 12:36:08,020::mem_free::53::mem_free.MemFree::(action ) memFree: 9816 Thread-3::INFO::2014-04-23 12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge ::(action) Found bridge ovirtmgmt Thread-296455::INFO::2014-04-23 12:36:08,678::listener::134::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(setup) Connection established Thread-296455::INFO::2014-04-23 12:36:08,684::listener::184::ovirt_hosted_engine _ha.broker.listener.ConnectionHandler::(handle) Connection closed /var/log/ovirt-hosted-engine-ha/agent.log host1: MainThread::INFO::2014-04-02 17:46:14,856::state_decorators::25::ovirt_hosted_en gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local engine vm statusno actions taken MainThread::INFO::2014-04-02 17:46:14,857::brokerlink::108::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1396453574.86 type=st ate_transition detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o virt.lan' MainThread::INFO::2014-04-02 17:46:14,858::brokerlink::117::ovirt_hosted_engine_ ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of state_transi tion (UnknownLocalVmState-UnknownLocalVmState) sent? ignored MainThread::WARNING::2014-04-02 17:46:15,463::hosted_engine::334::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error while monito ring engine: float() argument must be a string or a number MainThread::WARNING::2014-04-02 17:46:15,464::hosted_engine::337::ovirt_hosted_e ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Unexpected error Traceback (most recent call last): File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ ha/agent/hosted_eng ine.py, line 323, in start_monitoring state.score(self._log)) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ ha/agent/states.py , line 160, in score lm, logger, score, score_cfg) File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ ha/agent/states.py , line 61, in _penalize_memory if self._float_or_default(lm['mem-free'], 0) vm_mem: File /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ ha/agent/states.py , line 51, in _float_or_default return float(value) TypeError: float() argument must be a string or a number MainThread::ERROR::2014-04-02 17:46:15,464::hosted_engine::350::ovirt_hosted_eng ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Shutting down the ag ent because of 3 failures in a row! MainThread::INFO::2014-04-02 17:46:15,466::agent::116::ovirt_hosted_engine_ha.ag http://ovirt_hosted_engine_ha.ag ent.agent.Agent::(run) Agent shutting down host2: MainThread::INFO::2014-04-23 12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha. agent.hosted_engine.HostedEngine::(start_monitoring) Current state EngineUnexpectedlyDown (score: 0) MainThread::INFO::2014-04-23 12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib. brokerlink.BrokerLink::(notify) Trying: notify time=1398249414.84 type=state_transition detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown hostname='host02.ovirt.lan' MainThread::INFO::2014-04-23 12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib. brokerlink.BrokerLink::(notify) Success, was
[ovirt-users] oVirt Weekly Meeting Minutes -- 2014-04-23
Minutes: http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.html Minutes (text): http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.txt Log: http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.log.html = #ovirt: oVirt Weekly Sync = Meeting started by bkp at 14:04:45 UTC. The full logs are available at http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.log.html . Meeting summary --- * Agenda and roll Call (bkp, 14:05:24) * infra update (bkp, 14:05:24) * 3.3 updates (bkp, 14:05:24) * 3.4 status (bkp, 14:05:24) * 3.5 planning (bkp, 14:05:24) * conferences and workshops (bkp, 14:05:25) * other topics (bkp, 14:05:27) * infra update (bkp, 14:06:47) * OSAS sysadmin is working on updating MediaWiki on the oVirt site. (bkp, 14:09:53) * 3.3 updates (bkp, 14:10:27) * No 3.3 updates at this time. (bkp, 14:12:47) * 3.4 status (bkp, 14:13:11) * 3.4.1 composing is in process. Currently awaiting REST api packages. (doron, 14:22:46) * 3.5 status (doron, 14:23:11) * for next week please add missing devel owners for features planned for 3.5. (doron, 14:27:53) * infra 3.5 updates: several features merged, others in process more updates next week. Missing devel owners and bug should be completed by next week as well. (doron, 14:28:45) * 3.5 integration updates: Hosted engine support for ISCSI patch has been pushed. it seems to work but need heavy testing and reviewers. Guest Agent iso need upstream VMs with Windows, Lev need to sync with infra about that. (doron, 14:29:49) * 3.5 integration updates: Websocket on separate host, some progress there and DWH and reports on separate host may be at risk (doron, 14:30:31) * network 3.5 updates: some features are in review, others being discussed. We should re-evaluate next week. (doron, 14:33:50) * node 3.5 updates: appliance and hosted engine support doing well. Generic registration should be evaluated next week. (doron, 14:38:34) * sla 3.5 updates: most features doing good progress. specific ones will be re-evaluated next week. (doron, 14:39:50) * storage 3.5 updates: 2 features done. sanlock fencing at risk and the others are in progress. (doron, 14:45:49) * ux 3.5 updates: 'PatternFly adoption' is still in gerrit, pending review; it is not at risk for 3.5. 'GUI over rest-api' will not make it and should be removed. (doron, 14:47:35) * virt 3.5 updates: 6 features done. spice+vnc and instance types are currently at risk. Other features are in progress. (doron, 14:49:54) * for next week have every feature in every group ready for feature freeze review. (doron, 14:51:08) * conferences and workshops (doron, 14:51:21) * ovirt booth got lots of positive attention in redhat summit, more feedback to follow. (doron, 14:53:06) * upcoming events: LinuxFest Northwest this weekend. oVirt will have a shared booth (with Gluster and Atomic) there, and bkp is giving an oVirt talk. FISL 15 in Brazil the first week of May, as well as a separate CentOS dojo in Sao Paulo beforehand. (doron, 14:54:31) * other topics (doron, 14:56:23) * bkp looking for people interested in pitching in feedback/guidance on a site reboot for oVirt.org. We are thinking about moving away from MediaWiki to something with git-based source *and* easy-to-use editing. But that's just one way to go: feedback on design/infrastructure is very much needed. Email me if interested in joining the site committee. (doron, 14:58:57) * bkp expected to visit TLV site prior to openstack :) (doron, 14:59:26) Meeting ended at 15:02:03 UTC. Action Items Action Items, by person --- * **UNASSIGNED** * (none) People Present (lines said) --- * doron (86) * bkp (46) * lvernia (17) * fabiand (14) * mskrivanek (7) * ecohen (7) * ovedo (5) * ewoud (4) * amureini (4) * ybronhei (3) * ovirtbot (3) * knesenko (2) * xevilstar (2) * itamar (1) * karimb (1) * tiraboschi (1) Generated by `MeetBot`_ 0.1.4 .. _`MeetBot`: http://wiki.debian.org/MeetBot ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] hosted engine health check issues
same pb. ovirt-ha-broker have 400% cpu and is defunct. I can't kill with -9. 2014-04-23 13:55 GMT+02:00 Martin Sivak msi...@redhat.com: Hi, Isn't this file created when hosted engine is started? The file is created by the setup script. If it got lost then there was probably something bad happening in your NFS or Gluster storage. Or how can I create this file manually? I can give you experimental treatment for this. We do not have any official way as this is something that should not ever happen :) !! But before you do that make sure you do not have any nodes running properly. This will destroy and reinitialize the lockspace database for the whole hosted-engine environment (which you apparently lack, but..). !! You have to create the ha_agent/hosted-engine.lockspace file with the expected size (1MB) and then tell sanlock to initialize it as a lockspace using: # python import sanlock sanlock.write_lockspace(lockspace=hosted-engine, ... path=/rhev/data-center/mnt/nfs/hosted engine storage domain/ha_agent/hosted-engine.lockspace, ... offset=0) Then try starting the services (both broker and agent) again. -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 11:08 AM, Martin Sivak wrote: Hi René, libvirtError: Failed to acquire lock: No space left on device 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Can you please check the contents of /rhev/data-center/your nfs mount/nfs domain uuid/ha_agent/? This is how it should look like: [root@dev-03 ~]# ls -al /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/ total 2036 drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 . drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 .. -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata The errors seem to indicate that you somehow lost the lockspace file. True :) Isn't this file created when hosted engine is started? Or how can I create this file manually? -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 12:28 AM, Doron Fediuck wrote: Hi Rene, any idea what closed your ovirtmgmt bridge? as long as it is down vdsm may have issues starting up properly and this is why you see the complaints on the rpc server. Can you try manually fixing the network part first and then restart vdsm? Once vdsm is happy hosted engine VM will start. Thanks for your feedback, Doron. My ovirtmgmt bridge seems to be on or isn't it: # brctl show ovirtmgmt bridge namebridge id STP enabled interfaces ovirtmgmt 8000.0025907587c2 no eth0.200 # ip a s ovirtmgmt 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UNKNOWN link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever # ip a s eth0.200 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UP link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever I tried the following yesterday: Copy virtual disk from GlusterFS storage to local disk of host and create a new vm with virt-manager which loads ovirtmgmt disk. I could reach my engine over the ovirtmgmt bridge (so bridge must be working). I also started libvirtd with Option -v and I saw the following in libvirtd.log when trying to start ovirt engine: 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : Command result 0, with PID 11491 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is not a chain So it could be that something is broken in my hosted-engine network. Do you have any clue how I can troubleshoot this? Thanks, René - Original Message - From: René Koch rk...@linuxland.at To: Martin Sivak msi...@redhat.com Cc: users@ovirt.org Sent: Tuesday, April 22, 2014 1:46:38 PM Subject: Re: [ovirt-users] hosted engine health check issues Hi, I rebooted one of my ovirt hosts today and the result is now that I can't start hosted-engine anymore. ovirt-ha-agent isn't running because the lockspace file is missing (sanlock complains about it). So I tried to start hosted-engine with --vm-start and I get the
Re: [ovirt-users] hosted engine health check issues
Hi Kevin, same pb. Are you missing the lockspace file as well while running on top of GlusterFS? ovirt-ha-broker have 400% cpu and is defunct. I can't kill with -9. Defunct process eating full four cores? I wonder how is that possible.. What are the status flags of that process when you do ps axwu? Can you attach the log files please? -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - same pb. ovirt-ha-broker have 400% cpu and is defunct. I can't kill with -9. 2014-04-23 13:55 GMT+02:00 Martin Sivak msi...@redhat.com: Hi, Isn't this file created when hosted engine is started? The file is created by the setup script. If it got lost then there was probably something bad happening in your NFS or Gluster storage. Or how can I create this file manually? I can give you experimental treatment for this. We do not have any official way as this is something that should not ever happen :) !! But before you do that make sure you do not have any nodes running properly. This will destroy and reinitialize the lockspace database for the whole hosted-engine environment (which you apparently lack, but..). !! You have to create the ha_agent/hosted-engine.lockspace file with the expected size (1MB) and then tell sanlock to initialize it as a lockspace using: # python import sanlock sanlock.write_lockspace(lockspace=hosted-engine, ... path=/rhev/data-center/mnt/nfs/hosted engine storage domain/ha_agent/hosted-engine.lockspace, ... offset=0) Then try starting the services (both broker and agent) again. -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 11:08 AM, Martin Sivak wrote: Hi René, libvirtError: Failed to acquire lock: No space left on device 2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82 Can you please check the contents of /rhev/data-center/your nfs mount/nfs domain uuid/ha_agent/? This is how it should look like: [root@dev-03 ~]# ls -al /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/ total 2036 drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 . drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 .. -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata The errors seem to indicate that you somehow lost the lockspace file. True :) Isn't this file created when hosted engine is started? Or how can I create this file manually? -- Martin Sivák msi...@redhat.com Red Hat Czech RHEV-M SLA / Brno, CZ - Original Message - On 04/23/2014 12:28 AM, Doron Fediuck wrote: Hi Rene, any idea what closed your ovirtmgmt bridge? as long as it is down vdsm may have issues starting up properly and this is why you see the complaints on the rpc server. Can you try manually fixing the network part first and then restart vdsm? Once vdsm is happy hosted engine VM will start. Thanks for your feedback, Doron. My ovirtmgmt bridge seems to be on or isn't it: # brctl show ovirtmgmt bridge namebridge id STP enabled interfaces ovirtmgmt 8000.0025907587c2 no eth0.200 # ip a s ovirtmgmt 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UNKNOWN link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever # ip a s eth0.200 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue state UP link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff inet6 fe80::225:90ff:fe75:87c2/64 scope link valid_lft forever preferred_lft forever I tried the following yesterday: Copy virtual disk from GlusterFS storage to local disk of host and create a new vm with virt-manager which loads ovirtmgmt disk. I could reach my engine over the ovirtmgmt bridge (so bridge must be working). I also started libvirtd with Option -v and I saw the following in libvirtd.log when trying to start ovirt engine: 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : Command result 0, with PID 11491 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is not a chain So it could be that something is broken in my hosted-engine network. Do you have any clue how I can troubleshoot this?
Re: [ovirt-users] Ovirt snapshot failing on one VM
queries run on the system all the time but the only failure I encountered that was caused because of it was with live snapshots on a loaded setup in which vm running on the hsm would fail live snapshots if the system was loaded and the queries took a long time to come back. However, since the error you have happens when the vm is down and only on that specific vm, I think it's most likely related to a failure somewhere in the chain. Before committing or deleting any of the snapshots, is it possible for you to export the vm as is (with the snapshots) to an export domain? that way we know it's backed up before doing anything on the chain (and actually, this would be a much better way of backing up a vm rather than snapshots). I don't really know what the problem in the chain is or when it happened, which is why I want to be cautious when continuing and delete of a snapshot would be better than committing it. I can also suggest creating a new vm from any snapshot you think would be an important point in time for you - that case, even if there is a problem with the image you have a new vm with this image. so to recap, lets try this: 1. restart vdsm and try to create a snapshot again 2. export the vm to an export domain without collapsing the snapshots 3. delete or commit the snapshots - if any fail please attach the logs. also, if you delete/commit, after each snapshot you can try to create a new snapshot to see if the issue is solved On 04/23/2014 05:08 PM, Steve Dainard wrote: *Steve * On Wed, Apr 23, 2014 at 5:14 AM, Dafna Ron d...@redhat.com mailto:d...@redhat.com wrote: steve, I did not say that there is a limit. there is no limit and you can take a 1000 snapshots if you like, I simply said that I think that it would not be would a good practice to do so. I'm not trying to be adversarial here, but this is contradictory; if there's 'no limit' but 'its not good practice' and we assume that we want our virtual infrastructure to run smoothly, then effectively there is a limit we just don't know what it is. I also did not say that this is your current problem with the vm so you are jumping to conclusions here. I wasn't connecting the dots between # of snapshots, and the current issue, I have other VM's with the same amount of snapshots without this problem. No conclusion jumping going on. More interested in what the best practice is for VM's that accumulate snapshots over time. There is a feature slated for 3.5 http://www.ovirt.org/Features/Live_Merge which merges snapshots on a running VM, so I suppose in the long run I won't have a high snapshot count. i simply explained how snapshots work which is that they are created in a chain, if there is a problem at a single point in time it would effect the rest of the snapshots below it. Just for clarity, such a problem would affect the snapshots 'below it' means after the problematic snapshot? Example: Snapshot 1,2,3,4,5. #4 has a consistency issue, snaps 1,2,3 should be ok? I can try incrementally rolling back snapshots if this is the case (after vdsm restart suggested). Is there any way to do a consistency check? I can imagine scheduling a cronjob to run through a nightly check for consistency issues, then roll back to an earlier snapshot to circumvent the issue. And that we query all images under the base Image so if you have a lot of them it would take a long time for the results to come back. That's good to know, is this query done on new snapshot creation only? So over time the more snapshots I have, new snapshots will take longer to complete? as for your vm, since you fail to create a snapshot on only that vm it means that there is a problem in the current vm and it's chain. I can see when comparing the uuid's that the pool, domain, base image and last snapshots all exists in the rhev link. 2014-04-22 12:13:41,083 INFO [org.ovirt.engine.core.vdsbroker.irsbroker.CreateSnapshotVDSCommand] (pool-6-thread-49) [7ccaed5] -- createVolume parameters: sdUUID=95b9d922-4df7-4d3b-9bca-467e2fd9d573 spUUID=9497ef2c-8368-4c92-8d61-7f318a90748f imgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b size=21,474,836,480 bytes volFormat=COW volType=Sparse volUUID=0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936 descr= srcImgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b srcVolUUID=1a67de4b-aa1c-4436-baca-ca55726d54d7 lets see if it's possibly a cache issue - can you please restart vdsm on the hosts? I'll update when I have a chance to restart the services. Thanks On 04/22/2014 08:22 PM, Steve Dainard wrote: All snapshots are from before failure. That's a bit scary that there may be a 'too many snapshots'
[ovirt-users] problem in pulling zanata property files
Hi list, I'm in branch ovirt-engine-3.3, so I used the zanata-cli to pull property files from ovirt's zanata repo: https://translate.zanata.org/zanata/iteration/view/ovirt/ovirt-3.3 My steps: 1. download po files 2. config ~/.config/zanata.ini 3. config zanata.xml like this: ?xml version=1.0 encoding=UTF-8 standalone=yes? config xmlns=http://zanata.org/namespace/config/; urlhttps://translate.zanata.org/zanata//url projectovirt/project project-versionovirt-3.3/project-version !-- project-typeutf8properties|properties|gettext|podir|xliff|xml|file/project-type -- locales locale map-from=zh_CNzh-Hans-CN/locale /locales /config 4. execute pull -s {po files directory} -t {ovirt engine direstory} --project-type properties But after I use zanata-cli pull command, I get property files named with package name, like the following: $ git status # On branch ovirt-engine-3.3 # Changes not staged for commit: # (use git add file... to update what will be committed) # (use git checkout -- file... to discard changes in working directory) # # modified: frontend/webadmin/modules/uicompat/src/main/resources/org/ovirt/engine/ui/uicompat/LocalizedEnums_zh_CN.properties # modified: packaging/branding/ovirt.brand/messages_zh_CN.properties # # Untracked files: # (use git add file... to include in what will be committed) # # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.common.CommonApplicationConstants_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.common.CommonApplicationMessages_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.uicompat.SpiceRedKeys_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.uicompat.UIConstants_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.uicompat.UIMessages_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.userportal.ApplicationConstants_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.userportal.ApplicationMessages_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.webadmin.ApplicationConstants_zh_CN.properties # frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.webadmin.ApplicationMessages_zh_CN.properties no changes added to commit (use git add and/or git commit -a) I only see messages_zh_CN.properties and LocalizedEnums_zh_CN.properties are properly named, but not the rest property files. I noticed many translation po files under zanata project are all named with package name. So is this right ? thanks for your attention. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Anyone using gluster storage domain with WAN geo-rep?
I'm currently using a two node combined virt/storage setup with Ovirt 3.3.4 and Gluster 3.4.2 (replica 2, glusterfs storage domain). I'll call this pair PROD. I'm then geo-replicating to another gluster replica pair on the local net, btrfs underlying storage, and volume snapshots so I can recover my storage domain from different points in time if necessary. Its also local so restore time is much better than off-site. I'll call this pair BACKUP. I'm planning on setting up geo-replication from BACKUP to an EC2 gluster target. I'll call this host EC2HOST. PROD ---geo-rep-lan--- BACKUP ---geo-rep-wan--- EC2HOST I'd like to avoid saturating my WAN link during office hours. I have some ideas (or combination of): 1. limit bandwidth during certain hours to the offsite hosts. But realistically the bandwidth I would allocate is so low I don't see the purpose of this. Also with 8 guests running, I'm noticing quite a bit of data transfer to the local backup nodes (avg 6-8MB/s), and I'm thinking there is a lot of thrashing going on which isn't useful to backup offsite anyways. 2. stop WAN geo-replication during office hours, and restart for overnight/weekend hours. 3. Not use geo-rep between BACKUP --- EC2HOST, use rsync on one of the btrfs volume snapshots so we avoid the thrashing. In this case I could limit WAN speed to 1MB/s which should be fine for most differences throughout the day. So my question is, how do you off-site your storage domains, what constraints have you identified and how have you dealt with them? And of course how would you deal with the scenario I've oulined above? Thanks, *Steve* ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] qemu-kvm-rhev for el6
hello, does anyone know if there are an existent bugzilla to track the release of qemu-kvm-rhev rpms under el (like centos)? Because I've looked at bugzilla and google docs oVirt Planning Tracking with no luck best regards a ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] does SPM can run over ovirt-engine host ?
Hi , Yair yes, I dont want to re-install ovirt I'm not sure if allInone could be fix this problem. allinone install vdsm on the same ovirt-engine host. Well, I already have this : ovirt-engine, vdsm on the same host: srv-0202 My storage domains(data and iso) are hosted on srv-0202 I believe my solution is create one engine per server. Three independent engines managing only local virtual machines. On Mon, Apr 14, 2014 at 10:07 PM, Yair Zaslavsky yzasl...@redhat.comwrote: Hi Tamer, Are you familiar with the all in one feature? http://www.ovirt.org/Feature/AllInOne I'm not sure if this can help you now, as you probably don't want to re-install ovirt, right? - Original Message - From: Tamer Lima tamer.amer...@gmail.com To: users@ovirt.org Sent: Monday, April 14, 2014 5:13:12 PM Subject: [ovirt-users] does SPM can run over ovirt-engine host ? Hello, When I create virtual machine from a template (centos6.5, 2 cores, 8GB mem, 500GB hd) this process takes almost 2 hours. I click on New VM button and just select the template and click ok. engine.log show me high network consumption (98%) between engine-server host and SPM host. I tried to make my engine-server host a spm host too, but without sucess. Does SPM can run over on the same ovirt-engine machine ? Am I make something wrong? Or create VM from template is really slow ? my servers : srv-0202 = ovirt-engine , vdsm srv-0203 = spm , vdsm srv-0204 = vdsm These servers are dell blades connected on a 100GB switch. thanks This is what I know about SPM: http://www.ovirt.org/Storage_-_oVirt_workshop_November_2011 = Storage Pool Manager (SPM) A role assigned to one host in a data center granting it sole authority over: - Creation, deletion, an dmanipulation of virtula disk images, snapshots and templates - Templates: you can create on VM as a golden image and provision to multiple VMs (QCOW layers) - Allocation of storage for sparse block devices (on SAN) - Thin provisinoing (see below) - Single metadata writer: - SPM lease mechanism (Chockler and Malkhi 2004, Light-Weight Leases for Storage-Cnntric Coordination) - Storage-centric mailbox - This role can be migrated to any host in data center ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Foreman not starting VM, hangs on finishing creating a new VM
Hi All, I'm having an issue with oVirt 3.4 and Foreman 1.5 RC1 at the moment, it does not start/finish a new provisioned VM/Host., it creates it well on oVirt. I'm sure this worked well on a nightly some dags ago and I didn't update it after it. The foreman bug is shown here: http://projects.theforeman.org/issues/5132 What I also see in the engine log is a bunch of admin login errors which I cannot place anywhere: 2014-04-23 20:26:44,203 INFO [org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-6) Running command: LoginUserCommand internal: false. 2014-04-23 20:26:44,242 INFO [org.ovirt.engine.core.bll.LogoutUserCommand] (ajp--127.0.0.1-8702-6) [2e4a8183] Running command: LogoutUserCommand internal: false. 2014-04-23 20:26:44,276 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-6) [2e4a8183] Correlation ID: 2e4a8183, Call Stack: null, Custom Event ID: -1, Message: User admin logged out. 2014-04-23 20:26:44,470 INFO [org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-1) Running command: LoginUserCommand internal: false. 2014-04-23 20:26:44,510 INFO [org.ovirt.engine.core.bll.LogoutUserCommand] (ajp--127.0.0.1-8702-1) [5d731a43] Running command: LogoutUserCommand internal: false. 2014-04-23 20:26:44,542 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-1) [5d731a43] Correlation ID: 5d731a43, Call Stack: null, Custom Event ID: -1, Message: User admin logged out. 2014-04-23 20:26:44,575 INFO [org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-3) Running command: LoginUserCommand internal: false. 2014-04-23 20:26:44,623 INFO [org.ovirt.engine.core.bll.RemoveVmCommand] (ajp--127.0.0.1-8702-3) [708f0b9a] Lock Acquired to object EngineLock [exclusiveLocks= key: 7bfc175c-dbcd-43c9-9549-cde8d3b3b731 value: VM I get the feeling this has something todo with eachother. Any idea here ? Thanks! Matt ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Ovirt snapshot failing on one VM
On Wed, 23 Apr 2014, Steve Dainard wrote: I have other VM's with the same amount of snapshots without this problem. No conclusion jumping going on. More interested in what the best practice is for VM's that accumulate snapshots over time. For some real world context, we seem to accumulate snapshots using our local approach, and are not that focused on, or attentive about removing them. The 'highwater mark' of 39, on a machine that has been around since it was provisioned: 2010-01-05 [root@xxx backups]# ./count-snapshots.sh | sort -n | tail -3 38 vm_64099 38 vm_98036 39 vm_06359 Accumulating large numbers of snapshots seems more the function of pets, than ephemeral 'cattle' I wrote the first paragraph without looking up the 'owners' of the images. As I dereference the VM id's, all of the top ten in that list turn out to be mailservers, radius servers, name servers, and such, where the business unit owners chose not (or neglect) to 'winnow' their herd. There are no ephemeral use units in the top ten -- Russ herrold ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Unable to move disk to another data domain
- Original Message - From: Yusufi M R yu...@global-analytics.com To: users@ovirt.org Sent: Wednesday, April 23, 2014 5:22:43 PM Subject: [ovirt-users] Unable to move disk to another data domain Hi All, I am trying to move the disk of VM from Data Domain(Master) to another data domain in the same cluster, but I get the warning as “ some of the disk cannot be moved” and I cannot see the list of other data domain in the selection window. Can someone help here ? The message is displayed in case of unavailable source/target storage domains. Please verify that there's at least another one domain active in the data-center. Regards, Yusuf ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Unable to move disk to another data domain
- Original Message - From: Daniel Erez de...@redhat.com To: Yusufi M R yu...@global-analytics.com Cc: users@ovirt.org Sent: Wednesday, April 23, 2014 10:49:18 PM Subject: Re: [ovirt-users] Unable to move disk to another data domain - Original Message - From: Yusufi M R yu...@global-analytics.com To: users@ovirt.org Sent: Wednesday, April 23, 2014 5:22:43 PM Subject: [ovirt-users] Unable to move disk to another data domain Hi All, I am trying to move the disk of VM from Data Domain(Master) to another data domain in the same cluster, but I get the warning as “ some of the disk cannot be moved” and I cannot see the list of other data domain in the selection window. Can someone help here ? The message is displayed in case of unavailable source/target storage domains. Please verify that there's at least another one domain active in the data-center. Another possible issue could be that the disk's template is missing on the target domain, in this case you should first copy the template disk to the target domain. In case of live migration, currently the target domain should be of the same type as the source domain (you should have an indication of the exact problem with a tool-tip when hovering the blank target select-box). Regards, Yusuf ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Spurious error message when the RHEV capacity is maxed out?
We have about 900 vms in a RHEV-M 3.3 installation (3.3.2-0.50.el6ev) with three hosts. Each VM has 1 socket, 1 core for CPU and 256 MB memory and 2 x 1TB disks + one 5 GB boot disk. Each host has 2 cpu sockets with 4 cpu cores per socket. When we power on about 250 of these vms, we cannot power on any more vms. Attempting to do so gets the error (in the RHEV-M) console. Bad volume specification {'index': 0, 'iface': 'virtio', 'reqsize': '0', 'format': 'raw', 'bootOrder': '1', 'volumeID': '14aa7302-7f6d-4b80-922d-3fc090695447', 'apparentsize': '5368709120', 'imageID': '97b1985e-e507-4927-91e9-7393cde5', 'specParams': {}, 'readonly': 'false', 'domainID': '0c7abccc-1b97-4b39-ab96-e977a19fac3a', 'optional': 'false', 'deviceId': '97b1985e-e507-4927-91e9-7393cde5', 'truesize': '0', 'poolID': '5849b030-626e-47cb-ad90-3ce782d831b3', 'device': 'disk', 'shared': 'false', 'propagateErrors': 'off', 'type': 'disk'}. This looks like a spurious error, as we can power on the same VM after rebooting the hosts. And also, the hosts do not really look maxed out (CPU usage is around 25%, memory around 50%), but no more vm’s can be powered on. I was looking for a sizing guide, and found only https://access.redhat.com/site/sites/default/files/attachments/rhev_sizing_0812.pdf Is there any other information available? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] FC19 - Transaction check error - while installing ovirt-engine
I am seeing this on upgrade as well. So far I have not found a solution. On Mon 21 Apr 2014 12:48:20 AM PDT, Udaya Kiran P wrote: Hi, I am trying to install oVirt-Engine on Fedora-19. I am getting the below transaction check error while executing 'yum -y install ovirt-engine' command. Please help me resolve this. Transaction check error: file /usr/lib/python2.7/site-packages/sos/plugins/postgresql.py conflicts between attempted installs of ovirt-log-collector-3.4.1-1.fc19.noarch and sos-3.0-23.fc19.noarch file /usr/lib/python2.7/site-packages/sos/plugins/postgresql.pyc conflicts between attempted installs of ovirt-log-collector-3.4.1-1.fc19.noarch and sos-3.0-23.fc19.noarch file /usr/lib/python2.7/site-packages/sos/plugins/postgresql.pyo conflicts between attempted installs of ovirt-log-collector-3.4.1-1.fc19.noarch and sos-3.0-23.fc19.noarch I have used ovirt repo - yum localinstall http://resources.ovirt.org/releases/ovirt-release.noarch.rpm ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] FC19 - Transaction check error - while installing ovirt-engine
Ok, I managed to update with: yum --exclude=sos/* update I am seeing this on upgrade as well. So far I have not found a solution. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] qemu-kvm-rhev for el6
hello, does anyone know if there are an existent bugzilla to track the release of qemu-kvm-rhev rpms under el (like centos)? Because I've looked at bugzilla and google docs oVirt Planning Tracking with no luck best regards a I think this is the best fit that I have found so far: https://bugzilla.redhat.com/show_bug.cgi?id=1009100 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] configuration maximus
On 04/23/2014 11:49 PM, Ovirt User wrote: Hello, 1) how many Vm per node are supported ? should be ok with a couple of hundreds. 2) how many virtual CPU per node are supported ? virtual cpu are per VM? should be ok with 100 per VM (assuming you have a machine with enough cores!). I doubt anyone tested maxing these two figures out together, i don't see a relevant use case. i don't find configuration maximus .. thanks Lukas ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Spurious error message when the RHEV capacity is maxed out?
On 04/24/2014 12:31 AM, Satya Vempati wrote: We have about 900 vms in a RHEV-M 3.3 installation (3.3.2-0.50.el6ev) with three hosts. Each VM has 1 socket, 1 core for CPU and 256 MB memory and 2 x 1TB disks + one 5 GB boot disk. Each host has 2 cpu sockets with 4 cpu cores per socket. When we power on about 250 of these vms, we cannot power on any more vms. is this a 'monday morning effect' (are you trying to launch them all together)? Attempting to do so gets the error (in the RHEV-M) console. Bad volume specification {'index': 0, 'iface': 'virtio', 'reqsize': '0', 'format': 'raw', 'bootOrder': '1', 'volumeID': '14aa7302-7f6d-4b80-922d-3fc090695447', 'apparentsize': '5368709120', 'imageID': '97b1985e-e507-4927-91e9-7393cde5', 'specParams': {}, 'readonly': 'false', 'domainID': '0c7abccc-1b97-4b39-ab96-e977a19fac3a', 'optional': 'false', 'deviceId': '97b1985e-e507-4927-91e9-7393cde5', 'truesize': '0', 'poolID': '5849b030-626e-47cb-ad90-3ce782d831b3', 'device': 'disk', 'shared': 'false', 'propagateErrors': 'off', 'type': 'disk'}. This looks like a spurious error, as we can power on the same VM after rebooting the hosts. And also, the hosts do not really look maxed out (CPU usage is around 25%, memory around 50%), but no more vm’s can be powered on. I was looking for a sizing guide, and found only https://access.redhat.com/site/sites/default/files/attachments/rhev_sizing_0812.pdf Is there any other information available? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] FC19 - Transaction check error - while installing ovirt-engine
On 04/24/2014 03:21 AM, Thomas Suckow wrote: Ok, I managed to update with: yum --exclude=sos/* update I am seeing this on upgrade as well. So far I have not found a solution. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users is there a bug tracking this? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Foreman not starting VM, hangs on finishing creating a new VM
On 04/23/2014 09:50 PM, Matt . wrote: Hi All, I'm having an issue with oVirt 3.4 and Foreman 1.5 RC1 at the moment, it does not start/finish a new provisioned VM/Host., it creates it well on oVirt. I'm sure this worked well on a nightly some dags ago and I didn't update it after it. The foreman bug is shown here: http://projects.theforeman.org/issues/5132 What I also see in the engine log is a bunch of admin login errors which I cannot place anywhere: 2014-04-23 20:26:44,203 INFO [org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-6) Running command: LoginUserCommand internal: false. 2014-04-23 20:26:44,242 INFO [org.ovirt.engine.core.bll.LogoutUserCommand] (ajp--127.0.0.1-8702-6) [2e4a8183] Running command: LogoutUserCommand internal: false. 2014-04-23 20:26:44,276 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-6) [2e4a8183] Correlation ID: 2e4a8183, Call Stack: null, Custom Event ID: -1, Message: User admin logged out. 2014-04-23 20:26:44,470 INFO [org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-1) Running command: LoginUserCommand internal: false. 2014-04-23 20:26:44,510 INFO [org.ovirt.engine.core.bll.LogoutUserCommand] (ajp--127.0.0.1-8702-1) [5d731a43] Running command: LogoutUserCommand internal: false. 2014-04-23 20:26:44,542 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-1) [5d731a43] Correlation ID: 5d731a43, Call Stack: null, Custom Event ID: -1, Message: User admin logged out. 2014-04-23 20:26:44,575 INFO [org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-3) Running command: LoginUserCommand internal: false. 2014-04-23 20:26:44,623 INFO [org.ovirt.engine.core.bll.RemoveVmCommand] (ajp--127.0.0.1-8702-3) [708f0b9a] Lock Acquired to object EngineLock [exclusiveLocks= key: 7bfc175c-dbcd-43c9-9549-cde8d3b3b731 value: VM I get the feeling this has something todo with eachother. why do you think above snippet contain an error? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] does SPM can run over ovirt-engine host ?
On 04/23/2014 09:17 PM, Tamer Lima wrote: Hi , Yair yes, I dont want to re-install ovirt I'm not sure if allInone could be fix this problem. allinone install vdsm on the same ovirt-engine host. Well, I already have this : ovirt-engine, vdsm on the same host: srv-0202 My storage domains(data and iso) are hosted on srv-0202 I believe my solution is create one engine per server. Three independent engines managing only local virtual machines. this does not sounds right. engine and SPM should not communicate at 98% traffic for 2 hours. SPM should be one of the nodes in the DC. engine isn't acting as a node (even if deployed on one by chance of all-in-one or hosted engine). are you creating the VMs from template thinly provisioned or clone? On Mon, Apr 14, 2014 at 10:07 PM, Yair Zaslavsky yzasl...@redhat.com mailto:yzasl...@redhat.com wrote: Hi Tamer, Are you familiar with the all in one feature? http://www.ovirt.org/Feature/AllInOne I'm not sure if this can help you now, as you probably don't want to re-install ovirt, right? - Original Message - From: Tamer Lima tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com To: users@ovirt.org mailto:users@ovirt.org Sent: Monday, April 14, 2014 5:13:12 PM Subject: [ovirt-users] does SPM can run over ovirt-engine host ? Hello, When I create virtual machine from a template (centos6.5, 2 cores, 8GB mem, 500GB hd) this process takes almost 2 hours. I click on New VM button and just select the template and click ok. engine.log show me high network consumption (98%) between engine-server host and SPM host. I tried to make my engine-server host a spm host too, but without sucess. Does SPM can run over on the same ovirt-engine machine ? Am I make something wrong? Or create VM from template is really slow ? my servers : srv-0202 = ovirt-engine , vdsm srv-0203 = spm , vdsm srv-0204 = vdsm These servers are dell blades connected on a 100GB switch. thanks This is what I know about SPM: http://www.ovirt.org/Storage_-_oVirt_workshop_November_2011 = Storage Pool Manager (SPM) A role assigned to one host in a data center granting it sole authority over: - Creation, deletion, an dmanipulation of virtula disk images, snapshots and templates - Templates: you can create on VM as a golden image and provision to multiple VMs (QCOW layers) - Allocation of storage for sparse block devices (on SAN) - Thin provisinoing (see below) - Single metadata writer: - SPM lease mechanism (Chockler and Malkhi 2004, Light-Weight Leases for Storage-Cnntric Coordination) - Storage-centric mailbox - This role can be migrated to any host in data center ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users