Re: [ovirt-users] hosted engine health check issues

2014-04-23 Thread René Koch

On 04/23/2014 12:28 AM, Doron Fediuck wrote:

Hi Rene,
any idea what closed your ovirtmgmt bridge?
as long as it is down vdsm may have issues starting up properly
and this is why you see the complaints on the rpc server.

Can you try manually fixing the network part first and then
restart vdsm?
Once vdsm is happy hosted engine VM will start.


Thanks for your feedback, Doron.

My ovirtmgmt bridge seems to be on or isn't it:
# brctl show ovirtmgmt
bridge name bridge id   STP enabled interfaces
ovirtmgmt   8000.0025907587c2   no  eth0.200

# ip a s ovirtmgmt
7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue 
state UNKNOWN

link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt
inet6 fe80::225:90ff:fe75:87c2/64 scope link
   valid_lft forever preferred_lft forever

# ip a s eth0.200
6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc 
noqueue state UP

link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
inet6 fe80::225:90ff:fe75:87c2/64 scope link
   valid_lft forever preferred_lft forever

I tried the following yesterday:
Copy virtual disk from GlusterFS storage to local disk of host and 
create a new vm with virt-manager which loads ovirtmgmt disk. I could 
reach my engine over the ovirtmgmt bridge (so bridge must be working).


I also started libvirtd with Option -v and I saw the following in 
libvirtd.log when trying to start ovirt engine:
2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 : 
Command result 0, with PID 11491
2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result 
exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is 
not a chain


So it could be that something is broken in my hosted-engine network. Do 
you have any clue how I can troubleshoot this?



Thanks,
René




- Original Message -

From: René Koch rk...@linuxland.at
To: Martin Sivak msi...@redhat.com
Cc: users@ovirt.org
Sent: Tuesday, April 22, 2014 1:46:38 PM
Subject: Re: [ovirt-users] hosted engine health check issues

Hi,

I rebooted one of my ovirt hosts today and the result is now that I
can't start hosted-engine anymore.

ovirt-ha-agent isn't running because the lockspace file is missing
(sanlock complains about it).
So I tried to start hosted-engine with --vm-start and I get the
following errors:

== /var/log/sanlock.log ==
2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82

== /var/log/messages ==
Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 2014-04-22 12:38:17+0200 654
[3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name
2851af27-8744-445d-9fb1-a0d083c8dc82
Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering
disabled state
Apr 22 12:38:17 ovirt-host02 kernel: device vnet0 left promiscuous mode
Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering
disabled state

== /var/log/vdsm/vdsm.log ==
Thread-21::DEBUG::2014-04-22
12:38:17,563::libvirtconnection::124::root::(wrapper) Unknown
libvirterror: ecode: 38 edom: 42 level: 2 message: Failed to acquire
lock: No space left on device
Thread-21::DEBUG::2014-04-22
12:38:17,563::vm::2263::vm.Vm::(_startUnderlyingVm)
vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::_ongoingCreations released
Thread-21::ERROR::2014-04-22
12:38:17,564::vm::2289::vm.Vm::(_startUnderlyingVm)
vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed
Traceback (most recent call last):
File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm
  self._run()
File /usr/share/vdsm/vm.py, line 3170, in _run
  self._connection.createXML(domxml, flags),
File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py,
line 92, in wrapper
  ret = f(*args, **kwargs)
File /usr/lib64/python2.6/site-packages/libvirt.py, line 2665, in
createXML
  if ret is None:raise libvirtError('virDomainCreateXML() failed',
conn=self)
libvirtError: Failed to acquire lock: No space left on device

== /var/log/messages ==
Apr 22 12:38:17 ovirt-host02 vdsm vm.Vm ERROR
vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process
failed#012Traceback (most recent call last):#012  File
/usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm#012
self._run()#012  File /usr/share/vdsm/vm.py, line 3170, in _run#012
   self._connection.createXML(domxml, flags),#012  File
/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py, line 92,
in wrapper#012ret = f(*args, **kwargs)#012  File
/usr/lib64/python2.6/site-packages/libvirt.py, line 2665, in
createXML#012if ret is None:raise libvirtError('virDomainCreateXML()
failed', conn=self)#012libvirtError: Failed to acquire lock: No space
left on device

== /var/log/vdsm/vdsm.log ==
Thread-21::DEBUG::2014-04-22
12:38:17,569::vm::2731::vm.Vm::(setDownStatus)

Re: [ovirt-users] is spice html5 console actually working

2014-04-23 Thread David Jaša
Did you import your engine CA (https://ovirt-engine.example.org/ca.crt)
to firefox?

David

On Pá, 2014-04-18 at 16:07 -0500, Jeremiah Jahn wrote:
 I really seems to point to something strange that I've done with my
 certificates I guess..
 
 I get the following on chrome
 
 WebSocket server settings:
   - Listen on *:6100
   - Flash security policy server
   - SSL/TLS support
   - proxying from *:6100 to targets in /dummy
 
   1: handler exception: WSRequestHandler instance has no attribute 'last_code'
 
 
 and the following on firefox:
 
 [root@bluejay X86]#   - proxying from *:6100 to targets in /dummy
 
   1: handler exception: WSRequestHandler instance has no attribute 'last_code'
   2: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL
 routines:SSL3_READ_BYTES:sslv3 alert bad certificate
   3: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL
 routines:SSL3_READ_BYTES:sslv3 alert bad certificate
   4: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL
 routines:SSL3_READ_BYTES:tlsv1 alert access denied
   5: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL
 routines:SSL3_READ_BYTES:tlsv1 alert access denied
   6: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL
 routines:SSL3_READ_BYTES:tlsv1 alert access denied
 
 
 is there a way to just turn off SSL websockets for now?
 
 On Fri, Apr 18, 2014 at 9:24 AM, Jeremiah Jahn
 jerem...@goodinassociates.com wrote:
  Even after I've manually imported the ca.cert file into both chrome
  and firefox, I get the same error as before.  Is there something that
  needs to be done on each host?  I installed the websocket-proxy after
  the fact, and each host is a manually installed host, as opposed to
  the node.iso install.
 
  On Fri, Apr 18, 2014 at 8:28 AM, Jeremiah Jahn
  jerem...@goodinassociates.com wrote:
  And yet I always seem to get this error:
 
  WebSocket connection to
  'wss://localhost:6100/eyJ2YWxpZFRvIjoiMjAxNDA0MTgxMzI4MjMiLCJkYXRhIjoiJTdCJT...lHbjZVRHhzOThaRndVdUlmM0s5Wk1VQVh4K2QwZz09Iiwic2FsdCI6IjlxZFkzakN4U0lNPSJ9'
  failed: WebSocket is closed before the connection is established.
  spicehtml5-main.html?host=localhostport=6100:1
   disconnect spicehtml5-main.html?host=localhostport=6100:90
  Error in event handler for (unknown): TypeError: Cannot read property
  'state' of null at CSRecorder.onQueryStateCompleted
  (chrome-extension://cplklnmnlbnpmjogncfgfijoopmnlemp/content_scripts/recorder.js:43:13)
  at extensions::messaging:327:9 at Function.target.(anonymous function)
  (extensions::SafeBuiltins:19:14) at Event.dispatchToListener
  (extensions::event_bindings:386:22) at Event.dispatch_
  (extensions::event_bindings:371:27) at Event.dispatch
  (extensions::event_bindings:392:17) at dispatchOnMessage
  (extensions::messaging:294:22)
  spicehtml5-main.html?host=localhostport=6100:1
 
  On Thu, Apr 17, 2014 at 6:21 PM, Maurice James mja...@media-node.com 
  wrote:
  There are a few steps. Download the CA cert from your manager
  https://ovirtaddress/ca.crt
  Make sure it is trusted.
  Make sure ovirt-webproxy-socket is installed and running.
 
  Sent from my Galaxy S(R)III
 
 
   Original message 
  From: Jeremiah Jahn
  Date:04/17/2014 9:56 AM (GMT-05:00)
  To: users@ovirt.org
  Subject: [ovirt-users] is spice html5 console actually working
 
  Has anyone gotten the html5 spice console to work, and did you have to
  do anything special other than enable it?  I've tried every browser
  except opera and ie on linux and mac
  ___
  Users mailing list
  Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Feature Page: Mac Pool per DC

2014-04-23 Thread Martin Mucha
Hi, 

I was describing current state, first iteration. Need of restart is something 
which should not exist, I've removed that necessity meantime.
Altered flow: You allocate mac address for nic in data center without own pool, 
it gets registered in global pool. Then you modify settings of that data center 
so that new pool is created for it. All NICs for that data center is queries 
from DB, it's macs released from global pool and added to data center scope 
pool. And other way around. When you delete this scoped pool, all its content 
will be moved to global pool. Feature page is updated.

Note: *previously* there was MAC placed in wrong pool only after modification 
of existing data center, which caused entirely new pool to be created (there 
wasn't pool for this scope, after modification there is). All other operations 
were fine. Now all manipulation with scoped pools should be ok.

Note2: all that scoped pool handling is implemented as strategy. If we are 
unsatisfied with this implementation we could create another one and switch to 
it without modifying 'calling' code. Also many implementation may coexist and 
we can switch between them (on app start up) upon config.

Question: When allocating MAC, not one specified by user, system picks 
available mac from given mac pool. Imagine, that after some time then mac pool 
ranges changes, and lets say that whole new interval of macs is used, not 
overlapping with former one. Then all previously allocated macs will be present 
in altered pool as a user specified ones -- since they are outside of defined 
ranges. With large number of this mac address this have detrimental effect on 
memory usage. So if this is a real scenario, it would be acceptable(or 
welcomed) for you to reassign all mac address which were selected by system? 
For example on engine start / vm start.

M.

- Original Message -
From: Itamar Heim ih...@redhat.com
To: Martin Mucha mmu...@redhat.com
Cc: users@ovirt.org, de...@ovirt.org
Sent: Tuesday, April 22, 2014 5:15:35 PM
Subject: Re: [ovirt-users] Feature Page: Mac Pool per DC

On 04/18/2014 01:17 PM, Martin Mucha wrote:
 Hi,

 I'll try to describe it little bit more. Lets say, that we've got one data 
 center. It's not configured yet to have its own mac pool. So in system is 
 only one, global pool. We create few VMs and it's NICs will obtain its MAC 
 from this global pool, marking them as used. Next we alter data center 
 definition, so now it uses it's own mac pool. In system from this point on 
 exists two mac pools, one global and one related to this data center, but 
 those allocated MACs are still allocated in global pool, since new data 
 center creation does not (yet) contain logic to get all assigned MACs related 
 to this data center and reassign them in new pool. However, after app restart 
 all VmNics are read from db and placed to appropriate pools. Lets assume, 
 that we've performed such restart. Now we realized, that we actually don't 
 want that data center have own mac pool, so we alter it's definition removing 
 mac pool ranges. Pool related to this data center will be removed and it's 
 content will !
 be moved t
o a scope above this data center -- into global scope pool. We know, that 
everything what's allocated in pool to be removed is still used, but we need to 
track it elsewhere and currently there's just one option, global pool. So to 
answer your last question. When I remove scope, it's pool is gone and its 
content moved elsewhere. Next, when MAC is returned to the pool, the request 
goes like: give me pool for this virtual machine, and whatever pool it is, I'm 
returning this MAC to it. Clients of ScopedMacPoolManager do not know which 
pool they're talking to. Decision, which pool is right for them, is done behind 
the scenes upon their identification (I want pool for this logical network).

 Notice, that there is one problem in deciding which scope/pool to use. 
 There are places in code, which requires pool related to given data center, 
 identified by guid. For that request, only data center scope or something 
 broader like global scope can be returned. So even if one want to use one 
 pool per logical network, requests identified by data center id still can 
 return only data center scope or broader, and there are no chance returning 
 pool related to logical network (except for situation, where there is sole 
 logical network in that data center).

 Thanks for suggestion for another scopes. One question: if we're implementing 
 them, would you like just to pick a *sole* non-global scope you want to use 
 in your system (like data center related pools ONLY plus one global, or 
 logical network related pools ONLY plus one global) or would it be (more) 
 beneficial to you to have implemented some sort of cascading and overriding? 
 Like: this data center uses *this* pool, BUT except for *this* logical 
 network, which should use *this* one instead.

 I'll update feature page to contain these 

Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine

2014-04-23 Thread Dafna Ron
I think that you are mixing up a lot of different things and to be 
honest I am not sure what configuration you have and what exactly you 
are trying to do.

so lets try to simplify it?
what type of storage are you working on?
which host is the spm?


On 04/22/2014 07:36 PM, Tamer Lima wrote:

hello,

I am in trouble

I have 3 servers dedicated to test OVIRT:
01- engine + vdsm (8 cpus, 32GB ram , 2TB HD)
02 - vdsm (8 cpus, 32GB ram , 2TB HD)
03 - vdsm (8 cpus, 32GB ram , 2TB HD)

I want to create cloned virtual machines but in my configuration I can 
only save virtual machines on server 01; my configuration refers a 
DATA DOMAIN on server 01


All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were 
created like CLONE


My server 01 is the data domain and all new virtual machine is 
created, via NFS, on server 01 , who has 2TB maximum capacity ( the 
same size of partition /sda3 = 2TB)


how can I save each virtual machine on a desired vdsm server ?

What I want is :
server 01 - engine + vdsm : 03 virtual machines running and hosted 
phisicaly on this host
server 02 - vdsm : 04 virtual machines running and hosted phisicaly on 
this host
server 03 - vdsm : 04 virtual machines running and hosted phisicaly on 
this host


but I have this :
server 01 - engine + vdsm : 03 virtual machines running and hosted 
phisicaly on this host
server 02 - vdsm : 01 virtual machines running on this server BUT 
hosted phisicaly on server 01

server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB )

How to solve this problem ?
is it possible create one DATA DOMAIN for each VDSM host ? I think 
this is the solution but I do not know how to point VMs to be saved on 
specific data domain.


thanks




On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek 
michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote:



On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com wrote:

 hi, thanks for reply

 I am investigating what is and how thin virtualization works

 Do you know if HADOOP is indicated to work under thin
environment ?
 On Hadoop I will put large workloads and this thin
virtualization utilizes more resources than exists (shareable
environment)
 that is,
 if I have a real physical necessity of 500gb for each hadoop
host and my Thin Virtualization has 2TB on NFS, I can have only 4
virtual machines (500GB each), or less.

 For this case I believe clone virtual machine is the right
choice. But in my environment it takes 1h30m to build one cloned
virtual machine.

if you plan to overcommit then go with thin. The drawback is that
if you of course hit the physical limit the VMs will run out of
space...
if you plan to allocate 500GB each, consume all of it, never plan
to grow then go with the clone….yes, it's going to take time to
write all that stuff. With thin you need to do the same amount
of writes, but gradually over time while you're allocating it it

hope it helps

Thanks,
michal




 Am I correct ?





 On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek
michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com
wrote:

 On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com wrote:

 
 
  Hello,
 
  I created VMs by two ways :
 
  1) on tab virtual machines  new vm  template (centos_65_64bits)
  1.1 configuration : I do not select stateless checkbox
  1.2 this process takes a 1h30 to create each machine.
 
  2) on tab pools  new vm  template (centos_65_64bits)
  2.1 default configuration : stateless
  2.2 Here I created 3 virtual machines at once
  2.3 this process takes only one minute
 
  On the tab virtual machines I can see all virtual machines.
  Pooled machines have different icon image
  and description is different too:
 
  machines generated from tab VM are described as clone/dependent
  - clone is a phisical copy?
  machines generated from tab POOL are described as thin/independent
  - thin is a just a reference to template vm ? what is
phisical? any configuration file?

 yeah, sort of.
 just google thin provisioning in general:)


 
 
  In practice, what is the difference between these machines ?
 
 
 
 
  http://www.ovirt.org/Features/PrestartedVm
  Today there are 2 types of Vm pools:
  • Manual - the Vm is supposed to be manually returned to the
pool. In practice, this is not really entirely supported.
  • Automatic - once the user shuts down the Vm - it returns to
the pool (stateless).
 
  all vm created from pool are stateless ?

 the automatic pool, yes

 Thanks,
 michal

 
 
  thanks
 
 
 
  

Re: [ovirt-users] hosted engine health check issues

2014-04-23 Thread Martin Sivak
Hi René,

  libvirtError: Failed to acquire lock: No space left on device

  2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
  lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82

Can you please check the contents of /rhev/data-center/your nfs mount/nfs 
domain uuid/ha_agent/?

This is how it should look like:

[root@dev-03 ~]# ls -al 
/rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/
total 2036
drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 .
drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 ..
-rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace
-rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata

The errors seem to indicate that you somehow lost the lockspace file.

--
Martin Sivák
msi...@redhat.com
Red Hat Czech
RHEV-M SLA / Brno, CZ

- Original Message -
 On 04/23/2014 12:28 AM, Doron Fediuck wrote:
  Hi Rene,
  any idea what closed your ovirtmgmt bridge?
  as long as it is down vdsm may have issues starting up properly
  and this is why you see the complaints on the rpc server.
 
  Can you try manually fixing the network part first and then
  restart vdsm?
  Once vdsm is happy hosted engine VM will start.
 
 Thanks for your feedback, Doron.
 
 My ovirtmgmt bridge seems to be on or isn't it:
 # brctl show ovirtmgmt
 bridge name   bridge id   STP enabled interfaces
 ovirtmgmt 8000.0025907587c2   no  eth0.200
 
 # ip a s ovirtmgmt
 7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
 state UNKNOWN
  link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
  inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt
  inet6 fe80::225:90ff:fe75:87c2/64 scope link
 valid_lft forever preferred_lft forever
 
 # ip a s eth0.200
 6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc
 noqueue state UP
  link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
  inet6 fe80::225:90ff:fe75:87c2/64 scope link
 valid_lft forever preferred_lft forever
 
 I tried the following yesterday:
 Copy virtual disk from GlusterFS storage to local disk of host and
 create a new vm with virt-manager which loads ovirtmgmt disk. I could
 reach my engine over the ovirtmgmt bridge (so bridge must be working).
 
 I also started libvirtd with Option -v and I saw the following in
 libvirtd.log when trying to start ovirt engine:
 2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 :
 Command result 0, with PID 11491
 2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result
 exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is
 not a chain
 
 So it could be that something is broken in my hosted-engine network. Do
 you have any clue how I can troubleshoot this?
 
 
 Thanks,
 René
 
 
 
  - Original Message -
  From: René Koch rk...@linuxland.at
  To: Martin Sivak msi...@redhat.com
  Cc: users@ovirt.org
  Sent: Tuesday, April 22, 2014 1:46:38 PM
  Subject: Re: [ovirt-users] hosted engine health check issues
 
  Hi,
 
  I rebooted one of my ovirt hosts today and the result is now that I
  can't start hosted-engine anymore.
 
  ovirt-ha-agent isn't running because the lockspace file is missing
  (sanlock complains about it).
  So I tried to start hosted-engine with --vm-start and I get the
  following errors:
 
  == /var/log/sanlock.log ==
  2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
  lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82
 
  == /var/log/messages ==
  Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 2014-04-22 12:38:17+0200 654
  [3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name
  2851af27-8744-445d-9fb1-a0d083c8dc82
  Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering
  disabled state
  Apr 22 12:38:17 ovirt-host02 kernel: device vnet0 left promiscuous mode
  Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering
  disabled state
 
  == /var/log/vdsm/vdsm.log ==
  Thread-21::DEBUG::2014-04-22
  12:38:17,563::libvirtconnection::124::root::(wrapper) Unknown
  libvirterror: ecode: 38 edom: 42 level: 2 message: Failed to acquire
  lock: No space left on device
  Thread-21::DEBUG::2014-04-22
  12:38:17,563::vm::2263::vm.Vm::(_startUnderlyingVm)
  vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::_ongoingCreations released
  Thread-21::ERROR::2014-04-22
  12:38:17,564::vm::2289::vm.Vm::(_startUnderlyingVm)
  vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed
  Traceback (most recent call last):
  File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm
self._run()
  File /usr/share/vdsm/vm.py, line 3170, in _run
self._connection.createXML(domxml, flags),
  File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py,
  line 92, in wrapper
ret = f(*args, **kwargs)
  File 

Re: [ovirt-users] Feature Page: Mac Pool per DC

2014-04-23 Thread Sven Kieske
Hi,

I'm not 100% sure I understand the question.

What must be ensured: No VM, which got a
manually assigned address, which is (not)
in any configured pool, should stick to
this address.

I don't know if you plan to reassign
some random address from the same (or
any, if the mac is in no pool) pool.

But this should be avoided, as you may
rely on mac consistency for your vms.

Am 23.04.2014 10:12, schrieb Martin Mucha:
 So if this is a real scenario, it would be acceptable(or welcomed) for you to 
 reassign all mac address which were selected by system?

-- 
Mit freundlichen Grüßen / Regards

Sven Kieske

Systemadministrator
Mittwald CM Service GmbH  Co. KG
Königsberger Straße 6
32339 Espelkamp
T: +49-5772-293-100
F: +49-5772-293-333
https://www.mittwald.de
Geschäftsführer: Robert Meyer
St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen
Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted engine health check issues

2014-04-23 Thread René Koch

On 04/23/2014 11:08 AM, Martin Sivak wrote:

Hi René,


libvirtError: Failed to acquire lock: No space left on device



2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82


Can you please check the contents of /rhev/data-center/your nfs mount/nfs 
domain uuid/ha_agent/?

This is how it should look like:

[root@dev-03 ~]# ls -al 
/rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/
total 2036
drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 .
drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 ..
-rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace
-rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata

The errors seem to indicate that you somehow lost the lockspace file.


True :)
Isn't this file created when hosted engine is started? Or how can I 
create this file manually?




--
Martin Sivák
msi...@redhat.com
Red Hat Czech
RHEV-M SLA / Brno, CZ

- Original Message -

On 04/23/2014 12:28 AM, Doron Fediuck wrote:

Hi Rene,
any idea what closed your ovirtmgmt bridge?
as long as it is down vdsm may have issues starting up properly
and this is why you see the complaints on the rpc server.

Can you try manually fixing the network part first and then
restart vdsm?
Once vdsm is happy hosted engine VM will start.


Thanks for your feedback, Doron.

My ovirtmgmt bridge seems to be on or isn't it:
# brctl show ovirtmgmt
bridge name bridge id   STP enabled interfaces
ovirtmgmt   8000.0025907587c2   no  eth0.200

# ip a s ovirtmgmt
7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
state UNKNOWN
  link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
  inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt
  inet6 fe80::225:90ff:fe75:87c2/64 scope link
 valid_lft forever preferred_lft forever

# ip a s eth0.200
6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc
noqueue state UP
  link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
  inet6 fe80::225:90ff:fe75:87c2/64 scope link
 valid_lft forever preferred_lft forever

I tried the following yesterday:
Copy virtual disk from GlusterFS storage to local disk of host and
create a new vm with virt-manager which loads ovirtmgmt disk. I could
reach my engine over the ovirtmgmt bridge (so bridge must be working).

I also started libvirtd with Option -v and I saw the following in
libvirtd.log when trying to start ovirt engine:
2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 :
Command result 0, with PID 11491
2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result
exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is
not a chain

So it could be that something is broken in my hosted-engine network. Do
you have any clue how I can troubleshoot this?


Thanks,
René




- Original Message -

From: René Koch rk...@linuxland.at
To: Martin Sivak msi...@redhat.com
Cc: users@ovirt.org
Sent: Tuesday, April 22, 2014 1:46:38 PM
Subject: Re: [ovirt-users] hosted engine health check issues

Hi,

I rebooted one of my ovirt hosts today and the result is now that I
can't start hosted-engine anymore.

ovirt-ha-agent isn't running because the lockspace file is missing
(sanlock complains about it).
So I tried to start hosted-engine with --vm-start and I get the
following errors:

== /var/log/sanlock.log ==
2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82

== /var/log/messages ==
Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 2014-04-22 12:38:17+0200 654
[3093]: r2 cmd_acquire 2,9,5733 invalid lockspace found -1 failed 0 name
2851af27-8744-445d-9fb1-a0d083c8dc82
Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering
disabled state
Apr 22 12:38:17 ovirt-host02 kernel: device vnet0 left promiscuous mode
Apr 22 12:38:17 ovirt-host02 kernel: ovirtmgmt: port 2(vnet0) entering
disabled state

== /var/log/vdsm/vdsm.log ==
Thread-21::DEBUG::2014-04-22
12:38:17,563::libvirtconnection::124::root::(wrapper) Unknown
libvirterror: ecode: 38 edom: 42 level: 2 message: Failed to acquire
lock: No space left on device
Thread-21::DEBUG::2014-04-22
12:38:17,563::vm::2263::vm.Vm::(_startUnderlyingVm)
vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::_ongoingCreations released
Thread-21::ERROR::2014-04-22
12:38:17,564::vm::2289::vm.Vm::(_startUnderlyingVm)
vmId=`f26dd37e-13b5-430c-b2f2-ecd098b82a91`::The vm start process failed
Traceback (most recent call last):
 File /usr/share/vdsm/vm.py, line 2249, in _startUnderlyingVm
   self._run()
 File /usr/share/vdsm/vm.py, line 3170, in _run
   self._connection.createXML(domxml, flags),
 File /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py,
line 92, in wrapper
   ret = f(*args, **kwargs)
 File 

Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine

2014-04-23 Thread Sven Kieske
Hi,

the solution for this would be either
use gluster or another shared storage
software across all 3 hosts
or go with local storage datacenters.

Am 22.04.2014 20:36, schrieb Tamer Lima:
 hello,
 
 I am in trouble
 
 I have 3 servers dedicated to test OVIRT:
 01- engine  + vdsm   (8 cpus,  32GB ram , 2TB HD)
 02 - vdsm   (8 cpus,  32GB ram , 2TB HD)
 03 - vdsm   (8 cpus,  32GB ram , 2TB HD)
 
 I want to create cloned virtual machines but in my configuration I can only
 save virtual machines on server 01; my configuration refers a DATA DOMAIN
 on server 01
 
 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD  and were created
 like CLONE
 
 My server 01 is the data domain and all new virtual machine is created, via
 NFS,  on server 01 , who has 2TB maximum capacity   ( the same size of
 partition  /sda3 = 2TB)
 
 how can I save each virtual machine on a desired vdsm server ?
 
 What I want is :
 server 01 -   engine + vdsm  :03 virtual machines running and hosted
 phisicaly on this host
 server 02 -   vdsm  :04 virtual machines running and hosted  phisicaly
 on this host
 server 03 -   vdsm  :04 virtual machines running and hosted  phisicaly
 on this host
 
 but I have this :
 server 01 -   engine + vdsm  :03 virtual machines running  and hosted
 phisicaly on this host
 server 02 -   vdsm  :01 virtual machines running on this server  BUT
 hosted  phisicaly on server 01
 server 03 -   vdsm  :none, because my DATA DOMAIN IS FULL   (2TB )
 
 How to solve this problem ?
 is it possible create one DATA DOMAIN for each VDSM host  ?   I think this
 is the solution but I do not know how to point VMs to be saved on specific
 data domain.
 
 thanks
 
 
 
 
 On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek 
 michal.skriva...@redhat.com wrote:
 

 On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com wrote:

 hi,  thanks for reply

 I am investigating what is and how thin virtualization works

 Do you know if  HADOOP is indicated to work under thin environment ?
 On Hadoop I will put large workloads  and this  thin virtualization
  utilizes more resources than exists (shareable environment)
 that is,
 if I have a real physical necessity of 500gb for each hadoop host  and
 my Thin Virtualization has  2TB on NFS,  I can have only 4 virtual machines
  (500GB each), or less.

 For this case I believe clone virtual machine is the right choice. But
 in my environment it takes 1h30m to build one cloned virtual machine.

 if you plan to overcommit then go with thin. The drawback is that if you
 of course hit the physical limit the VMs will run out of space...
 if you plan to allocate 500GB each, consume all of it, never plan to grow
 then go with the clone….yes, it's going to take time to write all that
 stuff. With thin you need to do the same amount of writes, but gradually
 over time while you're allocating it it

 hope it helps

 Thanks,
 michal




 Am I correct ?





 On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek 
 michal.skriva...@redhat.com wrote:

 On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com wrote:



 Hello,

 I created VMs by two ways :

 1)  on tab virtual machines  new vm   template (centos_65_64bits)
 1.1 configuration : I do not select stateless checkbox
 1.2 this process takes a 1h30 to create each machine.

 2)  on tab pools  new vm   template (centos_65_64bits)
 2.1 default configuration : stateless
 2.2 Here I created 3 virtual machines at once
 2.3 this process takes only one minute

 On the tab virtual machines I can see all virtual machines.
 Pooled machines have different icon image
 and description is different too:

 machines generated from tab VM  are described as clone/dependent
 - clone is a phisical copy?
 machines generated from tab POOL are described as thin/independent
 - thin is a just a  reference to template vm ? what is phisical? any
 configuration file?

 yeah, sort of.
 just google thin provisioning in general:)




 In practice, what is the difference between these machines ?




 http://www.ovirt.org/Features/PrestartedVm
 Today there are 2 types of Vm pools:
   • Manual - the Vm is supposed to be manually returned to the
 pool. In practice, this is not really entirely supported.
   • Automatic - once the user shuts down the Vm - it returns to
 the pool (stateless).

  all vm created from pool  are stateless ?

 the automatic pool, yes

 Thanks,
 michal



 thanks



 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users




 
 
 
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users
 

-- 
Mit freundlichen Grüßen / Regards

Sven Kieske

Systemadministrator
Mittwald CM Service GmbH  Co. KG
Königsberger Straße 6
32339 Espelkamp
T: +49-5772-293-100
F: +49-5772-293-333
https://www.mittwald.de
Geschäftsführer: Robert Meyer
St.Nr.: 331/5721/1033, USt-IdNr.: 

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread Dafna Ron

steve,
I did not say that there is a limit. there is no limit and you can take 
a 1000 snapshots if you like, I simply said that I think that it would 
not be would a good practice to do so.
I also did not say that this is your current problem with the vm so you 
are jumping to conclusions here.
i simply explained how snapshots work which is that they are created in 
a chain, if there is a problem at a single point in time it would effect 
the rest of the snapshots below it.
And that we query all images under the base Image so if you have a lot 
of them it would take a long time for the results to come back.


as for your vm, since you fail to create a snapshot on only that vm it 
means that there is a problem in the current vm and it's chain.


I can see when comparing the uuid's that the pool, domain, base image 
and last snapshots all exists in the rhev link.


2014-04-22 12:13:41,083 INFO 
[org.ovirt.engine.core.vdsbroker.irsbroker.CreateSnapshotVDSCommand] 
(pool-6-thread-49) [7ccaed5] -- createVolume parameters:

sdUUID=95b9d922-4df7-4d3b-9bca-467e2fd9d573
spUUID=9497ef2c-8368-4c92-8d61-7f318a90748f
imgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
size=21,474,836,480 bytes
volFormat=COW
volType=Sparse
volUUID=0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936
descr=
srcImgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
srcVolUUID=1a67de4b-aa1c-4436-baca-ca55726d54d7



lets see if it's possibly a cache issue - can you please restart vdsm on 
the hosts?







On 04/22/2014 08:22 PM, Steve Dainard wrote:

All snapshots are from before failure.

That's a bit scary that there may be a 'too many snapshots' issue. I 
take snapshots for point in time consistency, and without the ability 
to collapse them while the vm is running I'm not sure what the best 
option is here. What is the recommended snapshot limit? Or maybe a 
better question; whats the intended use case for snapshots in ovirt?


Export domain is currently unavailable, and without it active I can't 
disable it properly.


# ls -tl 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b

total 8517740
-rw-rw. 1 vdsm kvm97583104 Apr 22 14:03 
1a67de4b-aa1c-4436-baca-ca55726d54d7
-rw-r--r--. 1 vdsm kvm 268 Apr 22 12:13 
1a67de4b-aa1c-4436-baca-ca55726d54d7.meta
-rw-r--r--. 1 vdsm kvm 272 Apr 22 01:06 
87390b64-becd-4a6f-a4fc-d27655f59b64.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 22 01:04 
1a67de4b-aa1c-4436-baca-ca55726d54d7.lease
-rw-rw. 1 vdsm kvm   107413504 Apr 20 22:00 
87390b64-becd-4a6f-a4fc-d27655f59b64
-rw-rw. 1 vdsm kvm   104267776 Apr 19 22:00 
6f9fd451-6c82-4390-802c-9e23a7d89427
-rw-rw. 1 vdsm kvm 1048576 Apr 19 22:00 
87390b64-becd-4a6f-a4fc-d27655f59b64.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 19 22:00 
6f9fd451-6c82-4390-802c-9e23a7d89427.meta
-rw-rw. 1 vdsm kvm   118358016 Apr 18 22:00 
c298ce3b-ec6a-4526-9971-a769f4d3d69b
-rw-rw. 1 vdsm kvm 1048576 Apr 18 22:00 
6f9fd451-6c82-4390-802c-9e23a7d89427.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 18 22:00 
c298ce3b-ec6a-4526-9971-a769f4d3d69b.meta
-rw-rw. 1 vdsm kvm   120913920 Apr 17 22:00 
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a
-rw-rw. 1 vdsm kvm 1048576 Apr 17 22:00 
c298ce3b-ec6a-4526-9971-a769f4d3d69b.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 17 22:00 
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.meta
-rw-rw. 1 vdsm kvm   117374976 Apr 16 22:00 
9aeb973d-9a54-441e-9ce9-f4f1a233da26
-rw-rw. 1 vdsm kvm 1048576 Apr 16 22:00 
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 16 22:00 
9aeb973d-9a54-441e-9ce9-f4f1a233da26.meta
-rw-rw. 1 vdsm kvm   110886912 Apr 15 22:00 
0eae2185-884a-44d3-9099-e952b6b7ec37
-rw-rw. 1 vdsm kvm 1048576 Apr 15 22:00 
9aeb973d-9a54-441e-9ce9-f4f1a233da26.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 15 22:00 
0eae2185-884a-44d3-9099-e952b6b7ec37.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 14 22:00 
0eae2185-884a-44d3-9099-e952b6b7ec37.lease
-rw-rw. 1 vdsm kvm   164560896 Apr 14 22:00 
ceffc643-b823-44b3-961e-93f3dc971886
-rw-r--r--. 1 vdsm kvm 272 Apr 14 22:00 
ceffc643-b823-44b3-961e-93f3dc971886.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 13 22:00 
ceffc643-b823-44b3-961e-93f3dc971886.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 13 22:00 
878fc690-ab08-489c-955b-9159f62026b1.meta
-rw-rw. 1 vdsm kvm   109182976 Apr 13 21:59 
878fc690-ab08-489c-955b-9159f62026b1
-rw-rw. 1 vdsm kvm   110297088 Apr 12 22:00 
5210eec2-a0eb-462e-95d5-7cf27db312f5
-rw-rw. 1 vdsm kvm 1048576 Apr 12 22:00 
878fc690-ab08-489c-955b-9159f62026b1.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 12 22:00 
5210eec2-a0eb-462e-95d5-7cf27db312f5.meta
-rw-rw. 1 vdsm kvm76480512 Apr 11 22:00 

Re: [ovirt-users] Feature Page: Mac Pool per DC

2014-04-23 Thread Sven Kieske
Sorry, I meant every vm
not No VM.

Am 23.04.2014 11:08, schrieb Sven Kieske:
 No VM, which got a
 manually assigned address, which is (not)
 in any configured pool, should stick to
 this address.

-- 
Mit freundlichen Grüßen / Regards

Sven Kieske

Systemadministrator
Mittwald CM Service GmbH  Co. KG
Königsberger Straße 6
32339 Espelkamp
T: +49-5772-293-100
F: +49-5772-293-333
https://www.mittwald.de
Geschäftsführer: Robert Meyer
St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen
Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine error -243

2014-04-23 Thread Kevin Tibi
Hi,

/var/log/ovirt-hosted-engine-ha/broker.log

Host1:
Thread-118327::INFO::2014-04-23
12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup)
Connection established
Thread-118327::INFO::2014-04-23
12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle)
Connection closed
Thread-118328::INFO::2014-04-23
12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup)
Connection established
Thread-118328::INFO::2014-04-23
12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle)
Connection closed

Host2:
Thread-4::INFO::2014-04-23
12:36:08,020::mem_free::53::mem_free.MemFree::(action)
memFree: 9816
Thread-3::INFO::2014-04-23
12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge
 ::(action) Found bridge ovirtmgmt
Thread-296455::INFO::2014-04-23
12:36:08,678::listener::134::ovirt_hosted_engine
 _ha.broker.listener.ConnectionHandler::(setup) Connection established
Thread-296455::INFO::2014-04-23
12:36:08,684::listener::184::ovirt_hosted_engine
 _ha.broker.listener.ConnectionHandler::(handle) Connection closed



/var/log/ovirt-hosted-engine-ha/agent.log

host1:

MainThread::INFO::2014-04-02
17:46:14,856::state_decorators::25::ovirt_hosted_en
  gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local engine vm
statusno actions taken
MainThread::INFO::2014-04-02
17:46:14,857::brokerlink::108::ovirt_hosted_engine_
  ha.lib.brokerlink.BrokerLink::(notify) Trying: notify time=1396453574.86
type=st   ate_transition
detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o
virt.lan'
MainThread::INFO::2014-04-02
17:46:14,858::brokerlink::117::ovirt_hosted_engine_
  ha.lib.brokerlink.BrokerLink::(notify) Success, was notification of
state_transi   tion
(UnknownLocalVmState-UnknownLocalVmState) sent? ignored
MainThread::WARNING::2014-04-02
17:46:15,463::hosted_engine::334::ovirt_hosted_e
ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error while
monito   ring engine: float() argument must be a
string or a number
MainThread::WARNING::2014-04-02
17:46:15,464::hosted_engine::337::ovirt_hosted_e
ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Unexpected
error
Traceback (most recent call last):
  File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/hosted_eng
ine.py, line 323, in start_monitoring
state.score(self._log))
  File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
, line 160, in score
lm, logger, score, score_cfg)
  File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
, line 61, in _penalize_memory
if self._float_or_default(lm['mem-free'], 0)  vm_mem:
  File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
, line 51, in _float_or_default
return float(value)
TypeError: float() argument must be a string or a number
MainThread::ERROR::2014-04-02
17:46:15,464::hosted_engine::350::ovirt_hosted_eng
  ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Shutting down
the ag   ent because of 3 failures in a row!
MainThread::INFO::2014-04-02 17:46:15,466::agent::116::
ovirt_hosted_engine_ha.ag   ent.agent.Agent::(run)
Agent shutting down


host2:

MainThread::INFO::2014-04-23
12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Current state EngineUnexpectedlyDown (score: 0)
MainThread::INFO::2014-04-23
12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1398249414.84 type=state_transition
detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown
hostname='host02.ovirt.lan'
MainThread::INFO::2014-04-23
12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Success, was notification of state_transition
(EngineUnexpectedlyDown-EngineUnexpectedlyDown) sent? ignored

/var/log/vdsm/vdsm.log

host1 :

Thread-116::DEBUG::2014-04-23
12:40:17,060::fileSD::225::Storage.Misc.excCmd::(getReadDelay) '/bin/dd
iflag=direct
if=/rhev/data-center/mnt/host01.ovirt.lan:_home_iso/cc51143e-8ad7-4b0b-a4d2-9024dffc1188/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-116::DEBUG::2014-04-23
12:40:17,070::fileSD::225::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
err = '0+1 records in\n0+1 records out\n343 bytes (343 B) copied,
0.000183642 s, 1.9 MB/s\n'; rc = 0
Thread-37::DEBUG::2014-04-23
12:40:17,504::fileSD::225::Storage.Misc.excCmd::(getReadDelay) '/bin/dd
iflag=direct
if=/rhev/data-center/mnt/host01.ovirt.lan:_home_NFS01/aea040f8-ab9d-435b-9ecf-ddd4272e592f/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-37::DEBUG::2014-04-23

Re: [ovirt-users] Hosted Engine error -243

2014-04-23 Thread Jiri Moskovcak

Hi,
I'm not sure yet what causes the problem, but the workaround should be:

open file 
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py 
in your favorite editor, go to line 52 and change it:


from: except ValueError:
to: except (ValueError, TypeError):

--Jirka

On 04/23/2014 12:43 PM, Kevin Tibi wrote:

Hi,

/var/log/ovirt-hosted-engine-ha/broker.log

Host1:
Thread-118327::INFO::2014-04-23
12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup)
Connection established
Thread-118327::INFO::2014-04-23
12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle)
Connection closed
Thread-118328::INFO::2014-04-23
12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(setup)
Connection established
Thread-118328::INFO::2014-04-23
12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker.listener.ConnectionHandler::(handle)
Connection closed

Host2:
Thread-4::INFO::2014-04-23
12:36:08,020::mem_free::53::mem_free.MemFree::(action
  ) memFree: 9816
Thread-3::INFO::2014-04-23
12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge
  ::(action) Found bridge ovirtmgmt
Thread-296455::INFO::2014-04-23
12:36:08,678::listener::134::ovirt_hosted_engine
  _ha.broker.listener.ConnectionHandler::(setup) Connection established
Thread-296455::INFO::2014-04-23
12:36:08,684::listener::184::ovirt_hosted_engine
  _ha.broker.listener.ConnectionHandler::(handle) Connection closed



/var/log/ovirt-hosted-engine-ha/agent.log

host1:

MainThread::INFO::2014-04-02
17:46:14,856::state_decorators::25::ovirt_hosted_en
   gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local
engine vm statusno actions taken
MainThread::INFO::2014-04-02
17:46:14,857::brokerlink::108::ovirt_hosted_engine_
   ha.lib.brokerlink.BrokerLink::(notify) Trying: notify
time=1396453574.86 type=st   ate_transition
detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o
 virt.lan'
MainThread::INFO::2014-04-02
17:46:14,858::brokerlink::117::ovirt_hosted_engine_
   ha.lib.brokerlink.BrokerLink::(notify) Success, was notification
of state_transi   tion
(UnknownLocalVmState-UnknownLocalVmState) sent? ignored
MainThread::WARNING::2014-04-02
17:46:15,463::hosted_engine::334::ovirt_hosted_e
   ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error
while monito   ring engine: float() argument
must be a string or a number
MainThread::WARNING::2014-04-02
17:46:15,464::hosted_engine::337::ovirt_hosted_e
   ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Unexpected error
Traceback (most recent call last):
   File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/hosted_eng
 ine.py, line 323, in start_monitoring
 state.score(self._log))
   File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
 , line 160, in score
 lm, logger, score, score_cfg)
   File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
 , line 61, in _penalize_memory
 if self._float_or_default(lm['mem-free'], 0)  vm_mem:
   File
/usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
 , line 51, in _float_or_default
 return float(value)
TypeError: float() argument must be a string or a number
MainThread::ERROR::2014-04-02
17:46:15,464::hosted_engine::350::ovirt_hosted_eng
 ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Shutting down the ag   ent because of 3 failures
in a row!
MainThread::INFO::2014-04-02
17:46:15,466::agent::116::ovirt_hosted_engine_ha.ag
http://ovirt_hosted_engine_ha.ag
ent.agent.Agent::(run) Agent shutting down


host2:

MainThread::INFO::2014-04-23
12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Current state EngineUnexpectedlyDown (score: 0)
MainThread::INFO::2014-04-23
12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1398249414.84 type=state_transition
detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown
hostname='host02.ovirt.lan'
MainThread::INFO::2014-04-23
12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Success, was notification of state_transition
(EngineUnexpectedlyDown-EngineUnexpectedlyDown) sent? ignored

/var/log/vdsm/vdsm.log

host1 :

Thread-116::DEBUG::2014-04-23
12:40:17,060::fileSD::225::Storage.Misc.excCmd::(getReadDelay) '/bin/dd
iflag=direct
if=/rhev/data-center/mnt/host01.ovirt.lan:_home_iso/cc51143e-8ad7-4b0b-a4d2-9024dffc1188/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-116::DEBUG::2014-04-23
12:40:17,070::fileSD::225::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
err = '0+1 

Re: [ovirt-users] hosted engine health check issues

2014-04-23 Thread Martin Sivak
Hi,

 Isn't this file created when hosted engine is started?

The file is created by the setup script. If it got lost then there was probably 
something bad happening in your NFS or Gluster storage.

 Or how can I create this file manually?

I can give you experimental treatment for this. We do not have any official way 
as this is something that should not ever happen :)

!! But before you do that make sure you do not have any nodes running properly. 
This will destroy and reinitialize the lockspace database for the whole 
hosted-engine environment (which you apparently lack, but..). !!

You have to create the ha_agent/hosted-engine.lockspace file with the expected 
size (1MB) and then tell sanlock to initialize it as a lockspace using:

# python
 import sanlock
 sanlock.write_lockspace(lockspace=hosted-engine,
... path=/rhev/data-center/mnt/nfs/hosted engine storage 
domain/ha_agent/hosted-engine.lockspace,
... offset=0)


Then try starting the services (both broker and agent) again.

--
Martin Sivák
msi...@redhat.com
Red Hat Czech
RHEV-M SLA / Brno, CZ


- Original Message -
 On 04/23/2014 11:08 AM, Martin Sivak wrote:
  Hi René,
 
  libvirtError: Failed to acquire lock: No space left on device
 
  2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
  lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82
 
  Can you please check the contents of /rhev/data-center/your nfs
  mount/nfs domain uuid/ha_agent/?
 
  This is how it should look like:
 
  [root@dev-03 ~]# ls -al
  /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/
  total 2036
  drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 .
  drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 ..
  -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace
  -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata
 
  The errors seem to indicate that you somehow lost the lockspace file.
 
 True :)
 Isn't this file created when hosted engine is started? Or how can I
 create this file manually?
 
 
  --
  Martin Sivák
  msi...@redhat.com
  Red Hat Czech
  RHEV-M SLA / Brno, CZ
 
  - Original Message -
  On 04/23/2014 12:28 AM, Doron Fediuck wrote:
  Hi Rene,
  any idea what closed your ovirtmgmt bridge?
  as long as it is down vdsm may have issues starting up properly
  and this is why you see the complaints on the rpc server.
 
  Can you try manually fixing the network part first and then
  restart vdsm?
  Once vdsm is happy hosted engine VM will start.
 
  Thanks for your feedback, Doron.
 
  My ovirtmgmt bridge seems to be on or isn't it:
  # brctl show ovirtmgmt
  bridge namebridge id   STP enabled interfaces
  ovirtmgmt  8000.0025907587c2   no  eth0.200
 
  # ip a s ovirtmgmt
  7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
  state UNKNOWN
link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt
inet6 fe80::225:90ff:fe75:87c2/64 scope link
   valid_lft forever preferred_lft forever
 
  # ip a s eth0.200
  6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc
  noqueue state UP
link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
inet6 fe80::225:90ff:fe75:87c2/64 scope link
   valid_lft forever preferred_lft forever
 
  I tried the following yesterday:
  Copy virtual disk from GlusterFS storage to local disk of host and
  create a new vm with virt-manager which loads ovirtmgmt disk. I could
  reach my engine over the ovirtmgmt bridge (so bridge must be working).
 
  I also started libvirtd with Option -v and I saw the following in
  libvirtd.log when trying to start ovirt engine:
  2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 :
  Command result 0, with PID 11491
  2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 : Result
  exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0' is
  not a chain
 
  So it could be that something is broken in my hosted-engine network. Do
  you have any clue how I can troubleshoot this?
 
 
  Thanks,
  René
 
 
 
  - Original Message -
  From: René Koch rk...@linuxland.at
  To: Martin Sivak msi...@redhat.com
  Cc: users@ovirt.org
  Sent: Tuesday, April 22, 2014 1:46:38 PM
  Subject: Re: [ovirt-users] hosted engine health check issues
 
  Hi,
 
  I rebooted one of my ovirt hosts today and the result is now that I
  can't start hosted-engine anymore.
 
  ovirt-ha-agent isn't running because the lockspace file is missing
  (sanlock complains about it).
  So I tried to start hosted-engine with --vm-start and I get the
  following errors:
 
  == /var/log/sanlock.log ==
  2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
  lockspace found -1 failed 0 name 2851af27-8744-445d-9fb1-a0d083c8dc82
 
  == /var/log/messages ==
  Apr 22 12:38:17 ovirt-host02 sanlock[3079]: 

Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine

2014-04-23 Thread Tamer Lima
hello,
thanks for reply

my storage is NFS v3, defined on host 01.  My DATA-DOMAIN and ISO-DOMAIN
are hosted on host 01;
my SPM is located on host 03, I dont remember why. I tried to migrate SPM
to host 01 but is not possible. All creation of virtual machine starts on
server 01 (






On Wed, Apr 23, 2014 at 5:19 AM, Dafna Ron d...@redhat.com wrote:

 I think that you are mixing up a lot of different things and to be honest
 I am not sure what configuration you have and what exactly you are trying
 to do.
 so lets try to simplify it?
 what type of storage are you working on?
 which host is the spm?



 On 04/22/2014 07:36 PM, Tamer Lima wrote:

 hello,

 I am in trouble

 I have 3 servers dedicated to test OVIRT:
 01- engine + vdsm (8 cpus, 32GB ram , 2TB HD)
 02 - vdsm (8 cpus, 32GB ram , 2TB HD)
 03 - vdsm (8 cpus, 32GB ram , 2TB HD)

 I want to create cloned virtual machines but in my configuration I can
 only save virtual machines on server 01; my configuration refers a DATA
 DOMAIN on server 01

 All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and were
 created like CLONE

 My server 01 is the data domain and all new virtual machine is created,
 via NFS, on server 01 , who has 2TB maximum capacity ( the same size of
 partition /sda3 = 2TB)

 how can I save each virtual machine on a desired vdsm server ?

 What I want is :
 server 01 - engine + vdsm : 03 virtual machines running and hosted
 phisicaly on this host
 server 02 - vdsm : 04 virtual machines running and hosted phisicaly on
 this host
 server 03 - vdsm : 04 virtual machines running and hosted phisicaly on
 this host

 but I have this :
 server 01 - engine + vdsm : 03 virtual machines running and hosted
 phisicaly on this host
 server 02 - vdsm : 01 virtual machines running on this server BUT hosted
 phisicaly on server 01
 server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB )

 How to solve this problem ?
 is it possible create one DATA DOMAIN for each VDSM host ? I think this
 is the solution but I do not know how to point VMs to be saved on specific
 data domain.

 thanks




 On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek 
 michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com wrote:


 On Apr 17, 2014, at 16:43 , Tamer Lima tamer.amer...@gmail.com
 mailto:tamer.amer...@gmail.com wrote:

  hi, thanks for reply
 
  I am investigating what is and how thin virtualization works
 
  Do you know if HADOOP is indicated to work under thin
 environment ?
  On Hadoop I will put large workloads and this thin
 virtualization utilizes more resources than exists (shareable
 environment)
  that is,
  if I have a real physical necessity of 500gb for each hadoop
 host and my Thin Virtualization has 2TB on NFS, I can have only 4
 virtual machines (500GB each), or less.
 
  For this case I believe clone virtual machine is the right
 choice. But in my environment it takes 1h30m to build one cloned
 virtual machine.

 if you plan to overcommit then go with thin. The drawback is that
 if you of course hit the physical limit the VMs will run out of
 space...
 if you plan to allocate 500GB each, consume all of it, never plan
 to grow then go with the clone….yes, it's going to take time to
 write all that stuff. With thin you need to do the same amount
 of writes, but gradually over time while you're allocating it it

 hope it helps

 Thanks,
 michal

 
 
 
  Am I correct ?
 
 
 
 
 
  On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek
 michal.skriva...@redhat.com mailto:michal.skriva...@redhat.com

 wrote:
 
  On Apr 16, 2014, at 16:41 , Tamer Lima tamer.amer...@gmail.com
 mailto:tamer.amer...@gmail.com wrote:
 
  
  
   Hello,
  
   I created VMs by two ways :
  
   1) on tab virtual machines  new vm  template (centos_65_64bits)
   1.1 configuration : I do not select stateless checkbox
   1.2 this process takes a 1h30 to create each machine.
  
   2) on tab pools  new vm  template (centos_65_64bits)
   2.1 default configuration : stateless
   2.2 Here I created 3 virtual machines at once
   2.3 this process takes only one minute
  
   On the tab virtual machines I can see all virtual machines.
   Pooled machines have different icon image
   and description is different too:
  
   machines generated from tab VM are described as clone/dependent
   - clone is a phisical copy?
   machines generated from tab POOL are described as thin/independent
   - thin is a just a reference to template vm ? what is
 phisical? any configuration file?
 
  yeah, sort of.
  just google thin provisioning in general:)
 
 
  
  
   In practice, what is the difference between these machines ?
  
  
  
  
   

Re: [ovirt-users] is spice html5 console actually working

2014-04-23 Thread Jeremiah Jahn
yes, yes I did...  It's as if installing the webproxy after the fact
caused it to self sign with it's own certificate or something.  I
edited the two html files so that for the moment it's not using https
and that works fine, but I just can't get the https to work at all.

On Wed, Apr 23, 2014 at 2:51 AM, David Jaša dj...@redhat.com wrote:
 Did you import your engine CA (https://ovirt-engine.example.org/ca.crt)
 to firefox?

 David

 On Pá, 2014-04-18 at 16:07 -0500, Jeremiah Jahn wrote:
 I really seems to point to something strange that I've done with my
 certificates I guess..

 I get the following on chrome

 WebSocket server settings:
   - Listen on *:6100
   - Flash security policy server
   - SSL/TLS support
   - proxying from *:6100 to targets in /dummy

   1: handler exception: WSRequestHandler instance has no attribute 
 'last_code'


 and the following on firefox:

 [root@bluejay X86]#   - proxying from *:6100 to targets in /dummy

   1: handler exception: WSRequestHandler instance has no attribute 
 'last_code'
   2: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL
 routines:SSL3_READ_BYTES:sslv3 alert bad certificate
   3: handler exception: [Errno 1] _ssl.c:1390: error:14094412:SSL
 routines:SSL3_READ_BYTES:sslv3 alert bad certificate
   4: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL
 routines:SSL3_READ_BYTES:tlsv1 alert access denied
   5: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL
 routines:SSL3_READ_BYTES:tlsv1 alert access denied
   6: handler exception: [Errno 1] _ssl.c:1390: error:14094419:SSL
 routines:SSL3_READ_BYTES:tlsv1 alert access denied


 is there a way to just turn off SSL websockets for now?

 On Fri, Apr 18, 2014 at 9:24 AM, Jeremiah Jahn
 jerem...@goodinassociates.com wrote:
  Even after I've manually imported the ca.cert file into both chrome
  and firefox, I get the same error as before.  Is there something that
  needs to be done on each host?  I installed the websocket-proxy after
  the fact, and each host is a manually installed host, as opposed to
  the node.iso install.
 
  On Fri, Apr 18, 2014 at 8:28 AM, Jeremiah Jahn
  jerem...@goodinassociates.com wrote:
  And yet I always seem to get this error:
 
  WebSocket connection to
  'wss://localhost:6100/eyJ2YWxpZFRvIjoiMjAxNDA0MTgxMzI4MjMiLCJkYXRhIjoiJTdCJT...lHbjZVRHhzOThaRndVdUlmM0s5Wk1VQVh4K2QwZz09Iiwic2FsdCI6IjlxZFkzakN4U0lNPSJ9'
  failed: WebSocket is closed before the connection is established.
  spicehtml5-main.html?host=localhostport=6100:1
   disconnect spicehtml5-main.html?host=localhostport=6100:90
  Error in event handler for (unknown): TypeError: Cannot read property
  'state' of null at CSRecorder.onQueryStateCompleted
  (chrome-extension://cplklnmnlbnpmjogncfgfijoopmnlemp/content_scripts/recorder.js:43:13)
  at extensions::messaging:327:9 at Function.target.(anonymous function)
  (extensions::SafeBuiltins:19:14) at Event.dispatchToListener
  (extensions::event_bindings:386:22) at Event.dispatch_
  (extensions::event_bindings:371:27) at Event.dispatch
  (extensions::event_bindings:392:17) at dispatchOnMessage
  (extensions::messaging:294:22)
  spicehtml5-main.html?host=localhostport=6100:1
 
  On Thu, Apr 17, 2014 at 6:21 PM, Maurice James mja...@media-node.com 
  wrote:
  There are a few steps. Download the CA cert from your manager
  https://ovirtaddress/ca.crt
  Make sure it is trusted.
  Make sure ovirt-webproxy-socket is installed and running.
 
  Sent from my Galaxy S(R)III
 
 
   Original message 
  From: Jeremiah Jahn
  Date:04/17/2014 9:56 AM (GMT-05:00)
  To: users@ovirt.org
  Subject: [ovirt-users] is spice html5 console actually working
 
  Has anyone gotten the html5 spice console to work, and did you have to
  do anything special other than enable it?  I've tried every browser
  except opera and ie on linux and mac
  ___
  Users mailing list
  Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine

2014-04-23 Thread Dafna Ron

what do you mean that host1 is engine + vdsm, are you using hosted engine?



On 04/23/2014 01:59 PM, Tamer Lima wrote:


hello,
thanks for reply

my storage is NFS v3, defined on host 01.  My DATA-DOMAIN and 
ISO-DOMAIN are hosted on host 01;
my SPM is located on host 03, I dont remember why. I tried to migrate 
SPM to host 01 but is not possible. All creation of virtual machine 
starts on server 01 (







On Wed, Apr 23, 2014 at 5:19 AM, Dafna Ron d...@redhat.com 
mailto:d...@redhat.com wrote:


I think that you are mixing up a lot of different things and to be
honest I am not sure what configuration you have and what exactly
you are trying to do.
so lets try to simplify it?
what type of storage are you working on?
which host is the spm?



On 04/22/2014 07:36 PM, Tamer Lima wrote:

hello,

I am in trouble

I have 3 servers dedicated to test OVIRT:
01- engine + vdsm (8 cpus, 32GB ram , 2TB HD)
02 - vdsm (8 cpus, 32GB ram , 2TB HD)
03 - vdsm (8 cpus, 32GB ram , 2TB HD)

I want to create cloned virtual machines but in my
configuration I can only save virtual machines on server 01;
my configuration refers a DATA DOMAIN on server 01

All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and
were created like CLONE

My server 01 is the data domain and all new virtual machine is
created, via NFS, on server 01 , who has 2TB maximum capacity
( the same size of partition /sda3 = 2TB)

how can I save each virtual machine on a desired vdsm server ?

What I want is :
server 01 - engine + vdsm : 03 virtual machines running and
hosted phisicaly on this host
server 02 - vdsm : 04 virtual machines running and hosted
phisicaly on this host
server 03 - vdsm : 04 virtual machines running and hosted
phisicaly on this host

but I have this :
server 01 - engine + vdsm : 03 virtual machines running and
hosted phisicaly on this host
server 02 - vdsm : 01 virtual machines running on this server
BUT hosted phisicaly on server 01
server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB )

How to solve this problem ?
is it possible create one DATA DOMAIN for each VDSM host ? I
think this is the solution but I do not know how to point VMs
to be saved on specific data domain.

thanks




On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek
michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com wrote:


On Apr 17, 2014, at 16:43 , Tamer Lima
tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com wrote:

 hi, thanks for reply

 I am investigating what is and how thin virtualization
works

 Do you know if HADOOP is indicated to work under thin
environment ?
 On Hadoop I will put large workloads and this thin
virtualization utilizes more resources than exists (shareable
environment)
 that is,
 if I have a real physical necessity of 500gb for each hadoop
host and my Thin Virtualization has 2TB on NFS, I can have
only 4
virtual machines (500GB each), or less.

 For this case I believe clone virtual machine is the right
choice. But in my environment it takes 1h30m to build one
cloned
virtual machine.

if you plan to overcommit then go with thin. The drawback
is that
if you of course hit the physical limit the VMs will run
out of
space...
if you plan to allocate 500GB each, consume all of it,
never plan
to grow then go with the clone….yes, it's going to take
time to
write all that stuff. With thin you need to do the same
amount
of writes, but gradually over time while you're allocating
it it

hope it helps

Thanks,
michal




 Am I correct ?





 On Thu, Apr 17, 2014 at 7:33 AM, Michal Skrivanek
michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com

wrote:

 On Apr 16, 2014, at 16:41 , Tamer Lima
tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com wrote:

 

[ovirt-users] Add a Direct Lun via rest API (Oivrt 3.3.5)

2014-04-23 Thread Gary Lloyd
Hello

I was just wondering if anyone would be able to help me figure out if there
is a way to login to an ISCSI target (EqualLogic) and add its associated
volume as a Direct LUN via the REST api.

I have figured out how to add an existing Direct LUN to a vm.

I have created a volume on the SAN and then I am attempting to upload some
xml to the API:

curl -v -u 'admin@internal:mypass' -H Content-type: application/xml -d
@disk.xml https://ovirt-test/disks/ --insecure

cat disk.xml

disk
aliasdirect_lun/alias
interfacevirtio/interface
formatraw/format
lunStorage
typeiscsi/type
logical_unit
address10.0.0.1/address
port3260/port

targetiqn.2001-05.com.equallogic:0-1cb196-cff1c713e-e2a004dfcc65357b-dev-directlun/target
/logical_unit
/lunStorage
/disk


At the moment the API is returning with a HTTP 400:

fault
reasonIncomplete parameters/reason
detailDisk [provisionedSize|size] required for add/detail
/fault

Is it possible to achieve my goal via the API ?

Thanks


*Gary Lloyd*
--
IT Services
Keele University
---
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] difference between thin/depentend and clone/dependent vm virtual machine

2014-04-23 Thread Tamer.americo
hi,

no. is not hosted engine.

on server 01 I installed ovirt engine , vdsm, libvirt, postgres , etc using 
engine-setup .  In this machine I have 3 VMs, all running centos65

this server 01 has ovirt web admin (jboss) and there I see all hosts ( server 
01, 02 and 03) on the cluster. I can create and manage VM (thin and cloned) 




tamer


Em 23/04/2014, às 10:21, Dafna Ron d...@redhat.com escreveu:

 what do you mean that host1 is engine + vdsm, are you using hosted engine?
 
 
 
 On 04/23/2014 01:59 PM, Tamer Lima wrote:
 
 hello,
 thanks for reply
 
 my storage is NFS v3, defined on host 01.  My DATA-DOMAIN and ISO-DOMAIN are 
 hosted on host 01;
 my SPM is located on host 03, I dont remember why. I tried to migrate SPM to 
 host 01 but is not possible. All creation of virtual machine starts on 
 server 01 (
 
 
 
 
 
 
 On Wed, Apr 23, 2014 at 5:19 AM, Dafna Ron d...@redhat.com 
 mailto:d...@redhat.com wrote:
 
I think that you are mixing up a lot of different things and to be
honest I am not sure what configuration you have and what exactly
you are trying to do.
so lets try to simplify it?
what type of storage are you working on?
which host is the spm?
 
 
 
On 04/22/2014 07:36 PM, Tamer Lima wrote:
 
hello,
 
I am in trouble
 
I have 3 servers dedicated to test OVIRT:
01- engine + vdsm (8 cpus, 32GB ram , 2TB HD)
02 - vdsm (8 cpus, 32GB ram , 2TB HD)
03 - vdsm (8 cpus, 32GB ram , 2TB HD)
 
I want to create cloned virtual machines but in my
configuration I can only save virtual machines on server 01;
my configuration refers a DATA DOMAIN on server 01
 
All my virtual machines are : 2 cpu , 6 GB ram , 500gb HD and
were created like CLONE
 
My server 01 is the data domain and all new virtual machine is
created, via NFS, on server 01 , who has 2TB maximum capacity
( the same size of partition /sda3 = 2TB)
 
how can I save each virtual machine on a desired vdsm server ?
 
What I want is :
server 01 - engine + vdsm : 03 virtual machines running and
hosted phisicaly on this host
server 02 - vdsm : 04 virtual machines running and hosted
phisicaly on this host
server 03 - vdsm : 04 virtual machines running and hosted
phisicaly on this host
 
but I have this :
server 01 - engine + vdsm : 03 virtual machines running and
hosted phisicaly on this host
server 02 - vdsm : 01 virtual machines running on this server
BUT hosted phisicaly on server 01
server 03 - vdsm : none, because my DATA DOMAIN IS FULL (2TB )
 
How to solve this problem ?
is it possible create one DATA DOMAIN for each VDSM host ? I
think this is the solution but I do not know how to point VMs
to be saved on specific data domain.
 
thanks
 
 
 
 
On Fri, Apr 18, 2014 at 4:48 AM, Michal Skrivanek
michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com
mailto:michal.skriva...@redhat.com wrote:
 
 
On Apr 17, 2014, at 16:43 , Tamer Lima
tamer.amer...@gmail.com mailto:tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com wrote:
 
 hi, thanks for reply

 I am investigating what is and how thin virtualization
works

 Do you know if HADOOP is indicated to work under thin
environment ?
 On Hadoop I will put large workloads and this thin
virtualization utilizes more resources than exists (shareable
environment)
 that is,
 if I have a real physical necessity of 500gb for each hadoop
host and my Thin Virtualization has 2TB on NFS, I can have
only 4
virtual machines (500GB each), or less.

 For this case I believe clone virtual machine is the right
choice. But in my environment it takes 1h30m to build one
cloned
virtual machine.
 
if you plan to overcommit then go with thin. The drawback
is that
if you of course hit the physical limit the VMs will run
out of
space...
if you plan to allocate 500GB each, consume all of it,
never plan
to grow then go with the clone….yes, it's going to take
time to
write all that stuff. With thin you need to do the same
amount
of writes, but gradually over time while you're allocating
it it
 
hope it helps
 
Thanks,
michal
 



 Am I correct ?





 On Thu, Apr 17, 2014 at 7:33 AM, 

[ovirt-users] Unable to move disk to another data domain

2014-04-23 Thread Yusufi M R
Hi All,

I am trying to move the disk of VM from Data Domain(Master) to another data 
domain in the same cluster, but I get the warning as  some of the disk cannot 
be moved  and I cannot see the list of other data domain in the selection 
window.

Can someone help here ?

Regards,
Yusuf
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine error -243

2014-04-23 Thread Kevin Tibi
In engine, i have
Hosted Engine HA: not activefor my host1
Hosted Engine HA: active (score 0)   for my host2




2014-04-23 13:52 GMT+02:00 Jiri Moskovcak jmosk...@redhat.com:

 Hi,
 I'm not sure yet what causes the problem, but the workaround should be:

 open file 
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
 in your favorite editor, go to line 52 and change it:

 from: except ValueError:
 to: except (ValueError, TypeError):

 --Jirka


 On 04/23/2014 12:43 PM, Kevin Tibi wrote:

 Hi,

 /var/log/ovirt-hosted-engine-ha/broker.log

 Host1:
 Thread-118327::INFO::2014-04-23
 12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(setup)
 Connection established
 Thread-118327::INFO::2014-04-23
 12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(handle)
 Connection closed
 Thread-118328::INFO::2014-04-23
 12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(setup)
 Connection established
 Thread-118328::INFO::2014-04-23
 12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(handle)
 Connection closed

 Host2:
 Thread-4::INFO::2014-04-23
 12:36:08,020::mem_free::53::mem_free.MemFree::(action
   ) memFree: 9816
 Thread-3::INFO::2014-04-23
 12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge
   ::(action) Found bridge ovirtmgmt
 Thread-296455::INFO::2014-04-23
 12:36:08,678::listener::134::ovirt_hosted_engine
   _ha.broker.listener.ConnectionHandler::(setup) Connection established
 Thread-296455::INFO::2014-04-23
 12:36:08,684::listener::184::ovirt_hosted_engine
   _ha.broker.listener.ConnectionHandler::(handle) Connection closed



 /var/log/ovirt-hosted-engine-ha/agent.log

 host1:

 MainThread::INFO::2014-04-02
 17:46:14,856::state_decorators::25::ovirt_hosted_en
gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local
 engine vm statusno actions taken
 MainThread::INFO::2014-04-02
 17:46:14,857::brokerlink::108::ovirt_hosted_engine_
ha.lib.brokerlink.BrokerLink::(notify) Trying: notify
 time=1396453574.86 type=st   ate_transition
 detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o
  virt.lan'
 MainThread::INFO::2014-04-02
 17:46:14,858::brokerlink::117::ovirt_hosted_engine_
ha.lib.brokerlink.BrokerLink::(notify) Success, was notification
 of state_transi   tion
 (UnknownLocalVmState-UnknownLocalVmState) sent? ignored
 MainThread::WARNING::2014-04-02
 17:46:15,463::hosted_engine::334::ovirt_hosted_e
ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error
 while monito   ring engine: float() argument
 must be a string or a number
 MainThread::WARNING::2014-04-02
 17:46:15,464::hosted_engine::337::ovirt_hosted_e
ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
 Unexpected error
 Traceback (most recent call last):
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/hosted_eng
  ine.py, line 323, in start_monitoring
  state.score(self._log))
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
  , line 160, in score
  lm, logger, score, score_cfg)
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
  , line 61, in _penalize_memory
  if self._float_or_default(lm['mem-free'], 0)  vm_mem:
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
  , line 51, in _float_or_default
  return float(value)
 TypeError: float() argument must be a string or a number
 MainThread::ERROR::2014-04-02
 17:46:15,464::hosted_engine::350::ovirt_hosted_eng
  ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
 Shutting down the ag   ent because of 3 failures
 in a row!
 MainThread::INFO::2014-04-02
 17:46:15,466::agent::116::ovirt_hosted_engine_ha.ag
 http://ovirt_hosted_engine_ha.ag

 ent.agent.Agent::(run) Agent shutting down


 host2:

 MainThread::INFO::2014-04-23
 12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha.
 agent.hosted_engine.HostedEngine::(start_monitoring)
 Current state EngineUnexpectedlyDown (score: 0)
 MainThread::INFO::2014-04-23
 12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib.
 brokerlink.BrokerLink::(notify)
 Trying: notify time=1398249414.84 type=state_transition
 detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown
 hostname='host02.ovirt.lan'
 MainThread::INFO::2014-04-23
 12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib.
 brokerlink.BrokerLink::(notify)
 Success, was notification of state_transition
 (EngineUnexpectedlyDown-EngineUnexpectedlyDown) sent? ignored

 /var/log/vdsm/vdsm.log

 host1 :

 Thread-116::DEBUG::2014-04-23
 

Re: [ovirt-users] Hosted Engine error -243

2014-04-23 Thread Kevin Tibi
Ho god, My cpu usage is 80% on host1

1729 vdsm  20   0  762m  15m 2884 S 297.6  0.1  77:16.70 ovirt-ha-broker



2014-04-23 16:40 GMT+02:00 Kevin Tibi kevint...@hotmail.com:

 In engine, i have
 Hosted Engine HA: not activefor my host1
  Hosted Engine HA: active (score 0)   for my host2




 2014-04-23 13:52 GMT+02:00 Jiri Moskovcak jmosk...@redhat.com:

 Hi,
 I'm not sure yet what causes the problem, but the workaround should be:

 open file 
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_ha/agent/states.py
 in your favorite editor, go to line 52 and change it:

 from: except ValueError:
 to: except (ValueError, TypeError):

 --Jirka


 On 04/23/2014 12:43 PM, Kevin Tibi wrote:

 Hi,

 /var/log/ovirt-hosted-engine-ha/broker.log

 Host1:
 Thread-118327::INFO::2014-04-23
 12:34:59,360::listener::134::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(setup)
 Connection established
 Thread-118327::INFO::2014-04-23
 12:34:59,375::listener::184::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(handle)
 Connection closed
 Thread-118328::INFO::2014-04-23
 12:35:14,546::listener::134::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(setup)
 Connection established
 Thread-118328::INFO::2014-04-23
 12:35:14,549::listener::184::ovirt_hosted_engine_ha.broker.
 listener.ConnectionHandler::(handle)
 Connection closed

 Host2:
 Thread-4::INFO::2014-04-23
 12:36:08,020::mem_free::53::mem_free.MemFree::(action
   ) memFree: 9816
 Thread-3::INFO::2014-04-23
 12:36:08,240::mgmt_bridge::59::mgmt_bridge.MgmtBridge
   ::(action) Found bridge ovirtmgmt
 Thread-296455::INFO::2014-04-23
 12:36:08,678::listener::134::ovirt_hosted_engine
   _ha.broker.listener.ConnectionHandler::(setup) Connection established
 Thread-296455::INFO::2014-04-23
 12:36:08,684::listener::184::ovirt_hosted_engine
   _ha.broker.listener.ConnectionHandler::(handle) Connection closed



 /var/log/ovirt-hosted-engine-ha/agent.log

 host1:

 MainThread::INFO::2014-04-02
 17:46:14,856::state_decorators::25::ovirt_hosted_en
gine_ha.agent.hosted_engine.HostedEngine::(check) Unknown local
 engine vm statusno actions taken
 MainThread::INFO::2014-04-02
 17:46:14,857::brokerlink::108::ovirt_hosted_engine_
ha.lib.brokerlink.BrokerLink::(notify) Trying: notify
 time=1396453574.86 type=st   ate_transition
 detail=UnknownLocalVmState-UnknownLocalVmState hostname='host01.o
  virt.lan'
 MainThread::INFO::2014-04-02
 17:46:14,858::brokerlink::117::ovirt_hosted_engine_
ha.lib.brokerlink.BrokerLink::(notify) Success, was notification
 of state_transi   tion
 (UnknownLocalVmState-UnknownLocalVmState) sent? ignored
 MainThread::WARNING::2014-04-02
 17:46:15,463::hosted_engine::334::ovirt_hosted_e
ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring) Error
 while monito   ring engine: float() argument
 must be a string or a number
 MainThread::WARNING::2014-04-02
 17:46:15,464::hosted_engine::337::ovirt_hosted_e
ngine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
 Unexpected error
 Traceback (most recent call last):
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_
 ha/agent/hosted_eng
  ine.py, line 323, in start_monitoring
  state.score(self._log))
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_
 ha/agent/states.py
  , line 160, in score
  lm, logger, score, score_cfg)
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_
 ha/agent/states.py
  , line 61, in _penalize_memory
  if self._float_or_default(lm['mem-free'], 0)  vm_mem:
File
 /usr/lib/python2.6/site-packages/ovirt_hosted_engine_
 ha/agent/states.py
  , line 51, in _float_or_default
  return float(value)
 TypeError: float() argument must be a string or a number
 MainThread::ERROR::2014-04-02
 17:46:15,464::hosted_engine::350::ovirt_hosted_eng
  ine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
 Shutting down the ag   ent because of 3 failures
 in a row!
 MainThread::INFO::2014-04-02
 17:46:15,466::agent::116::ovirt_hosted_engine_ha.ag
 http://ovirt_hosted_engine_ha.ag

 ent.agent.Agent::(run) Agent shutting down


 host2:

 MainThread::INFO::2014-04-23
 12:36:44,800::hosted_engine::323::ovirt_hosted_engine_ha.
 agent.hosted_engine.HostedEngine::(start_monitoring)
 Current state EngineUnexpectedlyDown (score: 0)
 MainThread::INFO::2014-04-23
 12:36:54,844::brokerlink::108::ovirt_hosted_engine_ha.lib.
 brokerlink.BrokerLink::(notify)
 Trying: notify time=1398249414.84 type=state_transition
 detail=EngineUnexpectedlyDown-EngineUnexpectedlyDown
 hostname='host02.ovirt.lan'
 MainThread::INFO::2014-04-23
 12:36:54,846::brokerlink::117::ovirt_hosted_engine_ha.lib.
 brokerlink.BrokerLink::(notify)
 Success, was 

[ovirt-users] oVirt Weekly Meeting Minutes -- 2014-04-23

2014-04-23 Thread Doron Fediuck
Minutes:        http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.html
Minutes (text): http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.txt
Log:            
http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.log.html

=
#ovirt: oVirt Weekly Sync
=


Meeting started by bkp at 14:04:45 UTC. The full logs are available at
http://ovirt.org/meetings/ovirt/2014/ovirt.2014-04-23-14.04.log.html .



Meeting summary
---
* Agenda and roll Call  (bkp, 14:05:24)
  * infra update  (bkp, 14:05:24)
  * 3.3 updates  (bkp, 14:05:24)
  * 3.4 status  (bkp, 14:05:24)
  * 3.5 planning  (bkp, 14:05:24)
  * conferences and workshops  (bkp, 14:05:25)
  * other topics  (bkp, 14:05:27)

* infra update  (bkp, 14:06:47)
  * OSAS sysadmin is working on updating MediaWiki on the oVirt site.
    (bkp, 14:09:53)

* 3.3 updates  (bkp, 14:10:27)
  * No 3.3 updates at this time.  (bkp, 14:12:47)

* 3.4 status  (bkp, 14:13:11)
  * 3.4.1 composing is in process. Currently awaiting REST api packages.
    (doron, 14:22:46)

* 3.5 status  (doron, 14:23:11)
  * for next week please add missing devel owners for features planned
    for 3.5.  (doron, 14:27:53)
  * infra 3.5 updates: several features merged, others in process more
    updates next week. Missing devel owners and bug should be completed
    by next week as well.  (doron, 14:28:45)
  * 3.5 integration updates: Hosted engine support for ISCSI patch has
    been pushed. it seems to work but need heavy testing and reviewers.
    Guest Agent iso need upstream VMs with Windows, Lev need to sync
    with infra about that.  (doron, 14:29:49)
  * 3.5 integration updates: Websocket on separate host, some progress
    there and DWH and reports on separate host may be at risk  (doron,
    14:30:31)
  * network 3.5 updates: some features are in review, others being
    discussed. We should re-evaluate next week.  (doron, 14:33:50)
  * node 3.5 updates: appliance and hosted engine support doing well.
    Generic registration should be evaluated next week.  (doron,
    14:38:34)
  * sla 3.5 updates: most features doing good progress. specific ones
    will be re-evaluated next week.  (doron, 14:39:50)
  * storage 3.5 updates: 2 features done. sanlock fencing at risk and
    the others are in progress.  (doron, 14:45:49)
  * ux 3.5 updates: 'PatternFly adoption' is still in gerrit, pending
    review; it is not at risk for 3.5. 'GUI over rest-api'  will not
    make it and should be removed.  (doron, 14:47:35)
  * virt 3.5 updates: 6 features done.  spice+vnc and instance types are
    currently at risk. Other features are in progress.  (doron,
    14:49:54)
  * for next week have  every feature in every group ready for feature
    freeze review.  (doron, 14:51:08)

* conferences and workshops  (doron, 14:51:21)
  * ovirt booth got lots of positive attention in redhat summit, more
    feedback to follow.  (doron, 14:53:06)
  * upcoming events: LinuxFest Northwest this weekend. oVirt will have a
    shared booth (with Gluster and Atomic) there, and bkp is giving an
    oVirt talk.  FISL 15 in Brazil the first week of May, as well as a
    separate CentOS dojo in Sao Paulo beforehand.  (doron, 14:54:31)

* other topics  (doron, 14:56:23)
  * bkp looking for people interested in pitching in feedback/guidance
    on a site reboot for oVirt.org. We are thinking about moving away
    from MediaWiki to something with git-based source *and* easy-to-use
    editing. But that's just one way to go: feedback on
    design/infrastructure is very much needed. Email me if interested in
    joining the site committee.  (doron, 14:58:57)
  * bkp expected to visit TLV site prior to openstack :)  (doron,
    14:59:26)

Meeting ended at 15:02:03 UTC.




Action Items






Action Items, by person
---
* **UNASSIGNED**
  * (none)




People Present (lines said)
---
* doron (86)
* bkp (46)
* lvernia (17)
* fabiand (14)
* mskrivanek (7)
* ecohen (7)
* ovedo (5)
* ewoud (4)
* amureini (4)
* ybronhei (3)
* ovirtbot (3)
* knesenko (2)
* xevilstar (2)
* itamar (1)
* karimb (1)
* tiraboschi (1)




Generated by `MeetBot`_ 0.1.4

.. _`MeetBot`: http://wiki.debian.org/MeetBot
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted engine health check issues

2014-04-23 Thread Kevin Tibi
same pb. ovirt-ha-broker have 400% cpu and is defunct. I can't kill with -9.


2014-04-23 13:55 GMT+02:00 Martin Sivak msi...@redhat.com:

 Hi,

  Isn't this file created when hosted engine is started?

 The file is created by the setup script. If it got lost then there was
 probably something bad happening in your NFS or Gluster storage.

  Or how can I create this file manually?

 I can give you experimental treatment for this. We do not have any
 official way as this is something that should not ever happen :)

 !! But before you do that make sure you do not have any nodes running
 properly. This will destroy and reinitialize the lockspace database for the
 whole hosted-engine environment (which you apparently lack, but..). !!

 You have to create the ha_agent/hosted-engine.lockspace file with the
 expected size (1MB) and then tell sanlock to initialize it as a lockspace
 using:

 # python
  import sanlock
  sanlock.write_lockspace(lockspace=hosted-engine,
 ... path=/rhev/data-center/mnt/nfs/hosted engine storage
 domain/ha_agent/hosted-engine.lockspace,
 ... offset=0)
 

 Then try starting the services (both broker and agent) again.

 --
 Martin Sivák
 msi...@redhat.com
 Red Hat Czech
 RHEV-M SLA / Brno, CZ


 - Original Message -
  On 04/23/2014 11:08 AM, Martin Sivak wrote:
   Hi René,
  
   libvirtError: Failed to acquire lock: No space left on device
  
   2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
   lockspace found -1 failed 0 name
 2851af27-8744-445d-9fb1-a0d083c8dc82
  
   Can you please check the contents of /rhev/data-center/your nfs
   mount/nfs domain uuid/ha_agent/?
  
   This is how it should look like:
  
   [root@dev-03 ~]# ls -al
  
 /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/
   total 2036
   drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 .
   drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 ..
   -rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace
   -rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata
  
   The errors seem to indicate that you somehow lost the lockspace file.
 
  True :)
  Isn't this file created when hosted engine is started? Or how can I
  create this file manually?
 
  
   --
   Martin Sivák
   msi...@redhat.com
   Red Hat Czech
   RHEV-M SLA / Brno, CZ
  
   - Original Message -
   On 04/23/2014 12:28 AM, Doron Fediuck wrote:
   Hi Rene,
   any idea what closed your ovirtmgmt bridge?
   as long as it is down vdsm may have issues starting up properly
   and this is why you see the complaints on the rpc server.
  
   Can you try manually fixing the network part first and then
   restart vdsm?
   Once vdsm is happy hosted engine VM will start.
  
   Thanks for your feedback, Doron.
  
   My ovirtmgmt bridge seems to be on or isn't it:
   # brctl show ovirtmgmt
   bridge namebridge id   STP enabled interfaces
   ovirtmgmt  8000.0025907587c2   no  eth0.200
  
   # ip a s ovirtmgmt
   7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
   state UNKNOWN
 link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
 inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt
 inet6 fe80::225:90ff:fe75:87c2/64 scope link
valid_lft forever preferred_lft forever
  
   # ip a s eth0.200
   6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc
   noqueue state UP
 link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
 inet6 fe80::225:90ff:fe75:87c2/64 scope link
valid_lft forever preferred_lft forever
  
   I tried the following yesterday:
   Copy virtual disk from GlusterFS storage to local disk of host and
   create a new vm with virt-manager which loads ovirtmgmt disk. I could
   reach my engine over the ovirtmgmt bridge (so bridge must be working).
  
   I also started libvirtd with Option -v and I saw the following in
   libvirtd.log when trying to start ovirt engine:
   2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 :
   Command result 0, with PID 11491
   2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 :
 Result
   exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0'
 is
   not a chain
  
   So it could be that something is broken in my hosted-engine network.
 Do
   you have any clue how I can troubleshoot this?
  
  
   Thanks,
   René
  
  
  
   - Original Message -
   From: René Koch rk...@linuxland.at
   To: Martin Sivak msi...@redhat.com
   Cc: users@ovirt.org
   Sent: Tuesday, April 22, 2014 1:46:38 PM
   Subject: Re: [ovirt-users] hosted engine health check issues
  
   Hi,
  
   I rebooted one of my ovirt hosts today and the result is now that I
   can't start hosted-engine anymore.
  
   ovirt-ha-agent isn't running because the lockspace file is missing
   (sanlock complains about it).
   So I tried to start hosted-engine with --vm-start and I get the
   

Re: [ovirt-users] hosted engine health check issues

2014-04-23 Thread Martin Sivak
Hi Kevin,

 same pb.

Are you missing the lockspace file as well while running on top of GlusterFS?

 ovirt-ha-broker have 400% cpu and is defunct. I can't kill with -9.

Defunct process eating full four cores? I wonder how is that possible.. What 
are the status flags of that process when you do ps axwu?

Can you attach the log files please?

--
Martin Sivák
msi...@redhat.com
Red Hat Czech
RHEV-M SLA / Brno, CZ

- Original Message -
 same pb. ovirt-ha-broker have 400% cpu and is defunct. I can't kill with -9.
 
 
 2014-04-23 13:55 GMT+02:00 Martin Sivak msi...@redhat.com:
 
  Hi,
 
   Isn't this file created when hosted engine is started?
 
  The file is created by the setup script. If it got lost then there was
  probably something bad happening in your NFS or Gluster storage.
 
   Or how can I create this file manually?
 
  I can give you experimental treatment for this. We do not have any
  official way as this is something that should not ever happen :)
 
  !! But before you do that make sure you do not have any nodes running
  properly. This will destroy and reinitialize the lockspace database for the
  whole hosted-engine environment (which you apparently lack, but..). !!
 
  You have to create the ha_agent/hosted-engine.lockspace file with the
  expected size (1MB) and then tell sanlock to initialize it as a lockspace
  using:
 
  # python
   import sanlock
   sanlock.write_lockspace(lockspace=hosted-engine,
  ... path=/rhev/data-center/mnt/nfs/hosted engine storage
  domain/ha_agent/hosted-engine.lockspace,
  ... offset=0)
  
 
  Then try starting the services (both broker and agent) again.
 
  --
  Martin Sivák
  msi...@redhat.com
  Red Hat Czech
  RHEV-M SLA / Brno, CZ
 
 
  - Original Message -
   On 04/23/2014 11:08 AM, Martin Sivak wrote:
Hi René,
   
libvirtError: Failed to acquire lock: No space left on device
   
2014-04-22 12:38:17+0200 654 [3093]: r2 cmd_acquire 2,9,5733 invalid
lockspace found -1 failed 0 name
  2851af27-8744-445d-9fb1-a0d083c8dc82
   
Can you please check the contents of /rhev/data-center/your nfs
mount/nfs domain uuid/ha_agent/?
   
This is how it should look like:
   
[root@dev-03 ~]# ls -al
   
  /rhev/data-center/mnt/euryale\:_home_ovirt_he/e16de6a2-53f5-4ab3-95a3-255d08398824/ha_agent/
total 2036
drwxr-x---. 2 vdsm kvm4096 Mar 19 18:46 .
drwxr-xr-x. 6 vdsm kvm4096 Mar 19 18:46 ..
-rw-rw. 1 vdsm kvm 1048576 Apr 23 11:05 hosted-engine.lockspace
-rw-rw. 1 vdsm kvm 1028096 Mar 19 18:46 hosted-engine.metadata
   
The errors seem to indicate that you somehow lost the lockspace file.
  
   True :)
   Isn't this file created when hosted engine is started? Or how can I
   create this file manually?
  
   
--
Martin Sivák
msi...@redhat.com
Red Hat Czech
RHEV-M SLA / Brno, CZ
   
- Original Message -
On 04/23/2014 12:28 AM, Doron Fediuck wrote:
Hi Rene,
any idea what closed your ovirtmgmt bridge?
as long as it is down vdsm may have issues starting up properly
and this is why you see the complaints on the rpc server.
   
Can you try manually fixing the network part first and then
restart vdsm?
Once vdsm is happy hosted engine VM will start.
   
Thanks for your feedback, Doron.
   
My ovirtmgmt bridge seems to be on or isn't it:
# brctl show ovirtmgmt
bridge namebridge id   STP enabled interfaces
ovirtmgmt  8000.0025907587c2   no  eth0.200
   
# ip a s ovirtmgmt
7: ovirtmgmt: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
state UNKNOWN
  link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
  inet 10.0.200.102/24 brd 10.0.200.255 scope global ovirtmgmt
  inet6 fe80::225:90ff:fe75:87c2/64 scope link
 valid_lft forever preferred_lft forever
   
# ip a s eth0.200
6: eth0.200@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc
noqueue state UP
  link/ether 00:25:90:75:87:c2 brd ff:ff:ff:ff:ff:ff
  inet6 fe80::225:90ff:fe75:87c2/64 scope link
 valid_lft forever preferred_lft forever
   
I tried the following yesterday:
Copy virtual disk from GlusterFS storage to local disk of host and
create a new vm with virt-manager which loads ovirtmgmt disk. I could
reach my engine over the ovirtmgmt bridge (so bridge must be working).
   
I also started libvirtd with Option -v and I saw the following in
libvirtd.log when trying to start ovirt engine:
2014-04-22 14:18:25.432+: 8901: debug : virCommandRunAsync:2250 :
Command result 0, with PID 11491
2014-04-22 14:18:25.478+: 8901: debug : virCommandRun:2045 :
  Result
exit status 255, stdout: '' stderr: 'iptables v1.4.7: goto 'FO-vnet0'
  is
not a chain
   
So it could be that something is broken in my hosted-engine network.
  Do
you have any clue how I can troubleshoot this?
  

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread Dafna Ron
queries run on the system all the time but the only failure I 
encountered that was caused because of it was with live snapshots on a 
loaded setup in which vm running on the hsm would fail live snapshots if 
the system was loaded and the queries took a long time to come back.
However, since the error you have happens when the vm is down and only 
on that specific vm, I think it's most likely related to a failure 
somewhere in the chain.


Before committing or deleting any of the snapshots, is it possible for 
you to export the vm as is (with the snapshots) to an export domain? 
that way we know it's backed up before doing anything on the chain (and 
actually, this would be a much better way of backing up a vm rather than 
snapshots).


I don't really know what the problem in the chain is or when it 
happened, which is why I want to be cautious when continuing and delete 
of a snapshot would be better than committing it.


I can also suggest creating a new vm from any snapshot you think would 
be an important point in time for you - that case, even if there is a 
problem with the image you have a new vm with this image.


so to recap, lets try this:
1. restart vdsm and try to create a snapshot again
2. export the vm to an export domain without collapsing the snapshots
3. delete or commit the snapshots - if any fail please attach the logs. 
also, if you delete/commit, after each snapshot you can try to create a 
new snapshot to see if the issue is solved






On 04/23/2014 05:08 PM, Steve Dainard wrote:



*Steve *

On Wed, Apr 23, 2014 at 5:14 AM, Dafna Ron d...@redhat.com 
mailto:d...@redhat.com wrote:


steve,
I did not say that there is a limit. there is no limit and you can
take a 1000 snapshots if you like, I simply said that I think that
it would not be would a good practice to do so.


I'm not trying to be adversarial here, but this is contradictory; if 
there's 'no limit' but 'its not good practice' and we assume that we 
want our virtual infrastructure to run smoothly, then effectively 
there is a limit we just don't know what it is.


I also did not say that this is your current problem with the vm
so you are jumping to conclusions here.


I wasn't connecting the dots between # of snapshots, and the current 
issue, I have other VM's with the same amount of snapshots without 
this problem. No conclusion jumping going on. More interested in what 
the best practice is for VM's that accumulate snapshots over time.


There is a feature slated for 3.5 
http://www.ovirt.org/Features/Live_Merge which merges snapshots on a 
running VM, so I suppose in the long run I won't have a high snapshot 
count.


i simply explained how snapshots work which is that they are
created in a chain, if there is a problem at a single point in
time it would effect the rest of the snapshots below it.


Just for clarity, such a problem would affect the snapshots 'below it' 
means after the problematic snapshot? Example: Snapshot 1,2,3,4,5. #4 
has a consistency issue, snaps 1,2,3 should be ok? I can try 
incrementally rolling back snapshots if this is the case (after vdsm 
restart suggested).


Is there any way to do a consistency check? I can imagine scheduling a 
cronjob to run through a nightly check for consistency issues, then 
roll back to an earlier snapshot to circumvent the issue.


And that we query all images under the base Image so if you have a
lot of them it would take a long time for the results to come back.


That's good to know, is this query done on new snapshot creation only? 
So over time the more snapshots I have, new snapshots will take longer 
to complete?



as for your vm, since you fail to create a snapshot on only that
vm it means that there is a problem in the current vm and it's chain.

I can see when comparing the uuid's that the pool, domain, base
image and last snapshots all exists in the rhev link.

2014-04-22 12:13:41,083 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.CreateSnapshotVDSCommand]
(pool-6-thread-49) [7ccaed5] -- createVolume parameters:
sdUUID=95b9d922-4df7-4d3b-9bca-467e2fd9d573
spUUID=9497ef2c-8368-4c92-8d61-7f318a90748f
imgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
size=21,474,836,480 bytes
volFormat=COW
volType=Sparse
volUUID=0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936
descr=
srcImgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
srcVolUUID=1a67de4b-aa1c-4436-baca-ca55726d54d7



lets see if it's possibly a cache issue - can you please restart
vdsm on the hosts?


I'll update when I have a chance to restart the services.

Thanks








On 04/22/2014 08:22 PM, Steve Dainard wrote:

All snapshots are from before failure.

That's a bit scary that there may be a 'too many snapshots'

[ovirt-users] problem in pulling zanata property files

2014-04-23 Thread plysan
Hi list,

I'm in branch ovirt-engine-3.3, so I used the zanata-cli to pull property
files from ovirt's zanata repo:
https://translate.zanata.org/zanata/iteration/view/ovirt/ovirt-3.3
My steps:
1. download po files
2. config ~/.config/zanata.ini
3. config zanata.xml like this:

?xml version=1.0 encoding=UTF-8 standalone=yes?
config xmlns=http://zanata.org/namespace/config/;
  urlhttps://translate.zanata.org/zanata//url
  projectovirt/project
  project-versionovirt-3.3/project-version
  !--
project-typeutf8properties|properties|gettext|podir|xliff|xml|file/project-type
--

  locales
locale map-from=zh_CNzh-Hans-CN/locale
  /locales

/config

4. execute pull -s {po files directory} -t {ovirt engine direstory}
--project-type properties

But after I use zanata-cli pull command, I get property files named with
package name, like the following:

$ git status
# On branch ovirt-engine-3.3
# Changes not staged for commit:
#   (use git add file... to update what will be committed)
#   (use git checkout -- file... to discard changes in working
directory)
#
# modified:
frontend/webadmin/modules/uicompat/src/main/resources/org/ovirt/engine/ui/uicompat/LocalizedEnums_zh_CN.properties
# modified:   packaging/branding/ovirt.brand/messages_zh_CN.properties
#
# Untracked files:
#   (use git add file... to include in what will be committed)
#
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.common.CommonApplicationConstants_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.common.CommonApplicationMessages_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.uicompat.SpiceRedKeys_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.uicompat.UIConstants_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.uicompat.UIMessages_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.userportal.ApplicationConstants_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.userportal.ApplicationMessages_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.webadmin.ApplicationConstants_zh_CN.properties
#
frontend/webadmin/modules/webadmin/src/main/resources/org/ovirt/engine/ui/frontend/org.ovirt.engine.ui.webadmin.ApplicationMessages_zh_CN.properties
no changes added to commit (use git add and/or git commit -a)

I only see messages_zh_CN.properties and LocalizedEnums_zh_CN.properties
are properly named, but not the rest property files.
I noticed many translation po files under zanata project are all named with
package name.
So is this right ?

thanks for your attention.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Anyone using gluster storage domain with WAN geo-rep?

2014-04-23 Thread Steve Dainard
I'm currently using a two node combined virt/storage setup with Ovirt 3.3.4
and Gluster 3.4.2 (replica 2, glusterfs storage domain). I'll call this
pair PROD.

I'm then geo-replicating to another gluster replica pair on the local net,
btrfs underlying storage, and volume snapshots so I can recover my storage
domain from different points in time if necessary. Its also local so
restore time is much better than off-site. I'll call this pair BACKUP.

I'm planning on setting up geo-replication from BACKUP to an EC2 gluster
target. I'll call this host EC2HOST.

PROD ---geo-rep-lan--- BACKUP ---geo-rep-wan--- EC2HOST

I'd like to avoid saturating my WAN link during office hours. I have some
ideas (or combination of):

1. limit bandwidth during certain hours to the offsite hosts. But
realistically the bandwidth I would allocate is so low I don't see the
purpose of this. Also with 8 guests running, I'm noticing quite a bit of
data transfer to the local backup nodes (avg 6-8MB/s), and I'm thinking
there is a lot of thrashing going on which isn't useful to backup offsite
anyways.

2. stop WAN geo-replication during office hours, and restart for
overnight/weekend hours.

3. Not use geo-rep between BACKUP --- EC2HOST, use rsync on one of the
btrfs volume snapshots so we avoid the thrashing. In this case I could
limit WAN speed to 1MB/s which should be fine for most differences
throughout the day.

So my question is, how do you off-site your storage domains, what
constraints have you identified and how have you dealt with them? And of
course how would you deal with the scenario I've oulined above?

Thanks,




*Steve*
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] qemu-kvm-rhev for el6

2014-04-23 Thread Amedeo Salvati

hello,

does anyone know if there are an existent bugzilla to track the release 
of qemu-kvm-rhev rpms under el (like centos)?
Because I've looked at bugzilla and google docs oVirt Planning  
Tracking with no luck


best regards
a
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] does SPM can run over ovirt-engine host ?

2014-04-23 Thread Tamer Lima
Hi , Yair

yes, I dont want to re-install ovirt

I'm not sure if allInone could be fix this problem.

allinone install vdsm on the same ovirt-engine host. Well, I already have
this : ovirt-engine, vdsm on the same host: srv-0202

My storage domains(data and iso) are hosted on srv-0202


I believe my solution is create one engine per server. Three independent
engines managing only local virtual machines.






On Mon, Apr 14, 2014 at 10:07 PM, Yair Zaslavsky yzasl...@redhat.comwrote:

 Hi Tamer,
 Are you familiar with the all in one feature?

 http://www.ovirt.org/Feature/AllInOne

 I'm not sure if this can help you now, as you probably  don't want to
 re-install ovirt, right?


 - Original Message -
  From: Tamer Lima tamer.amer...@gmail.com
  To: users@ovirt.org
  Sent: Monday, April 14, 2014 5:13:12 PM
  Subject: [ovirt-users] does SPM can run over ovirt-engine host ?
 
  Hello,
 
  When I create virtual machine from a template (centos6.5, 2 cores, 8GB
 mem,
  500GB hd)  this process takes almost 2 hours.   I click on New VM
 button
  and just select the template and click ok.
 
  engine.log show me high network consumption (98%)  between  engine-server
  host and SPM host.
 
  I tried to make my engine-server host a spm host too, but without sucess.
 
 
  Does SPM can run over on the same ovirt-engine machine ?
 
  Am I make something wrong? Or create VM from template is really slow ?
 
 
  my servers :
  srv-0202  = ovirt-engine  , vdsm
  srv-0203 = spm , vdsm
  srv-0204 = vdsm
  These servers are dell blades connected on a 100GB switch.
 
 
 
  thanks
 
 
 
 
  This is what I know about SPM:
  http://www.ovirt.org/Storage_-_oVirt_workshop_November_2011
 
  = Storage Pool Manager (SPM) A role assigned to one host in a data center
  granting it sole authority over:
 
 - Creation, deletion, an dmanipulation of virtula disk images,
 snapshots
 and templates
- Templates: you can create on VM as a golden image and provision
 to
multiple VMs (QCOW layers)
 - Allocation of storage for sparse block devices (on SAN)
- Thin provisinoing (see below)
 - Single metadata writer:
- SPM lease mechanism (Chockler and Malkhi 2004, Light-Weight
 Leases
for Storage-Cnntric Coordination)
- Storage-centric mailbox
 - This role can be migrated to any host in data center
 
  ___
  Users mailing list
  Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Foreman not starting VM, hangs on finishing creating a new VM

2014-04-23 Thread Matt .
Hi All,

I'm having an issue with oVirt 3.4 and Foreman 1.5 RC1 at the moment, it
does not start/finish a new provisioned VM/Host., it creates it well on
oVirt.

I'm sure this worked well on a nightly some dags ago and I didn't update it
after it.

The foreman bug is shown here: http://projects.theforeman.org/issues/5132

What I also see in the engine log is a bunch of admin login errors which I
cannot place anywhere:

2014-04-23 20:26:44,203 INFO  [org.ovirt.engine.core.bll.LoginUserCommand]
(ajp--127.0.0.1-8702-6) Running command: LoginUserCommand internal: false.
2014-04-23 20:26:44,242 INFO  [org.ovirt.engine.core.bll.LogoutUserCommand]
(ajp--127.0.0.1-8702-6) [2e4a8183] Running command: LogoutUserCommand
internal: false.
2014-04-23 20:26:44,276 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-6) [2e4a8183] Correlation ID: 2e4a8183, Call Stack:
null, Custom Event ID: -1, Message: User admin logged out.
2014-04-23 20:26:44,470 INFO  [org.ovirt.engine.core.bll.LoginUserCommand]
(ajp--127.0.0.1-8702-1) Running command: LoginUserCommand internal: false.
2014-04-23 20:26:44,510 INFO  [org.ovirt.engine.core.bll.LogoutUserCommand]
(ajp--127.0.0.1-8702-1) [5d731a43] Running command: LogoutUserCommand
internal: false.
2014-04-23 20:26:44,542 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-1) [5d731a43] Correlation ID: 5d731a43, Call Stack:
null, Custom Event ID: -1, Message: User admin logged out.
2014-04-23 20:26:44,575 INFO  [org.ovirt.engine.core.bll.LoginUserCommand]
(ajp--127.0.0.1-8702-3) Running command: LoginUserCommand internal: false.
2014-04-23 20:26:44,623 INFO  [org.ovirt.engine.core.bll.RemoveVmCommand]
(ajp--127.0.0.1-8702-3) [708f0b9a] Lock Acquired to object EngineLock
[exclusiveLocks= key: 7bfc175c-dbcd-43c9-9549-cde8d3b3b731 value: VM

I get the feeling this has something todo with eachother.

Any idea here ?

Thanks!

Matt
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread R P Herrold
On Wed, 23 Apr 2014, Steve Dainard wrote:

 I have other VM's with the same amount of snapshots without this problem.
 No conclusion jumping going on. More interested in what the best practice
 is for VM's that accumulate snapshots over time.

For some real world context, we seem to accumulate snapshots 
using our local approach, and are not that focused on, or 
attentive about removing them.  The 'highwater mark' of 39, on 
a machine that has been around since it was provisioned: 
2010-01-05

[root@xxx backups]# ./count-snapshots.sh | sort -n | tail -3
38 vm_64099
38 vm_98036
39 vm_06359

Accumulating large numbers of snapshots seems more the 
function of pets, than ephemeral 'cattle'

I wrote the first paragraph without looking up the 'owners' of 
the images. As I dereference the VM id's, all of the top ten 
in that list turn out to be mailservers, radius servers, name 
servers, and such, where the business unit owners chose not 
(or neglect) to 'winnow' their herd.  There are no ephemeral 
use units in the top ten

-- Russ herrold
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Unable to move disk to another data domain

2014-04-23 Thread Daniel Erez


- Original Message -
 From: Yusufi M R yu...@global-analytics.com
 To: users@ovirt.org
 Sent: Wednesday, April 23, 2014 5:22:43 PM
 Subject: [ovirt-users] Unable to move disk to another data domain
 
 
 
 Hi All,
 
 
 
 I am trying to move the disk of VM from Data Domain(Master) to another data
 domain in the same cluster, but I get the warning as “ some of the disk
 cannot be moved” and I cannot see the list of other data domain in the
 selection window.
 
 
 
 Can someone help here ?

The message is displayed in case of unavailable source/target storage domains.
Please verify that there's at least another one domain active in the 
data-center.

 
 
 
 Regards,
 
 Yusuf
 
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users
 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Unable to move disk to another data domain

2014-04-23 Thread Daniel Erez


- Original Message -
 From: Daniel Erez de...@redhat.com
 To: Yusufi M R yu...@global-analytics.com
 Cc: users@ovirt.org
 Sent: Wednesday, April 23, 2014 10:49:18 PM
 Subject: Re: [ovirt-users] Unable to move disk to another data domain
 
 
 
 - Original Message -
  From: Yusufi M R yu...@global-analytics.com
  To: users@ovirt.org
  Sent: Wednesday, April 23, 2014 5:22:43 PM
  Subject: [ovirt-users] Unable to move disk to another data domain
  
  
  
  Hi All,
  
  
  
  I am trying to move the disk of VM from Data Domain(Master) to another data
  domain in the same cluster, but I get the warning as “ some of the disk
  cannot be moved” and I cannot see the list of other data domain in the
  selection window.
  
  
  
  Can someone help here ?
 
 The message is displayed in case of unavailable source/target storage
 domains.
 Please verify that there's at least another one domain active in the
 data-center.

Another possible issue could be that the disk's template is missing
on the target domain, in this case you should first copy the template disk
to the target domain. In case of live migration, currently the target domain
should be of the same type as the source domain (you should have an indication
of the exact problem with a tool-tip when hovering the blank target select-box).

 
  
  
  
  Regards,
  
  Yusuf
  
  ___
  Users mailing list
  Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
  
 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Spurious error message when the RHEV capacity is maxed out?

2014-04-23 Thread Satya Vempati
We have about 900 vms in a RHEV-M 3.3 installation (3.3.2-0.50.el6ev) with 
three hosts.

Each VM has 1 socket, 1 core for CPU and 256 MB memory and 2 x 1TB disks + one 
5 GB boot disk.

Each host has 2 cpu sockets with 4 cpu cores per socket.

When we power on about 250 of these vms, we cannot power on any more vms.

Attempting to do so gets the error (in the RHEV-M) console.

Bad volume specification {'index': 0, 'iface': 'virtio', 'reqsize': '0', 
'format': 'raw', 'bootOrder': '1', 'volumeID': 
'14aa7302-7f6d-4b80-922d-3fc090695447', 'apparentsize': '5368709120', 
'imageID': '97b1985e-e507-4927-91e9-7393cde5', 'specParams': {}, 
'readonly': 'false', 'domainID': '0c7abccc-1b97-4b39-ab96-e977a19fac3a', 
'optional': 'false', 'deviceId': '97b1985e-e507-4927-91e9-7393cde5', 
'truesize': '0', 'poolID': '5849b030-626e-47cb-ad90-3ce782d831b3', 'device': 
'disk', 'shared': 'false', 'propagateErrors': 'off', 'type': 'disk'}.

This looks like a spurious error, as we can power on the same VM after 
rebooting the hosts.

And also, the hosts do not really look maxed out (CPU usage is around 25%, 
memory around 50%), but no more vm’s can be powered on.

I was looking for a sizing guide, and found only 
https://access.redhat.com/site/sites/default/files/attachments/rhev_sizing_0812.pdf

Is there any other information available?

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] FC19 - Transaction check error - while installing ovirt-engine

2014-04-23 Thread Thomas Suckow

I am seeing this on upgrade as well. So far I have not found a solution.

On Mon 21 Apr 2014 12:48:20 AM PDT, Udaya Kiran P wrote:

Hi,

I am trying to install oVirt-Engine on Fedora-19. I am getting the below 
transaction check error while executing 'yum -y install ovirt-engine' command.

Please help me resolve this.

Transaction check error:
  file /usr/lib/python2.7/site-packages/sos/plugins/postgresql.py conflicts 
between attempted installs of ovirt-log-collector-3.4.1-1.fc19.noarch and 
sos-3.0-23.fc19.noarch
  file /usr/lib/python2.7/site-packages/sos/plugins/postgresql.pyc conflicts 
between attempted installs of ovirt-log-collector-3.4.1-1.fc19.noarch and 
sos-3.0-23.fc19.noarch
  file /usr/lib/python2.7/site-packages/sos/plugins/postgresql.pyo conflicts 
between attempted installs of ovirt-log-collector-3.4.1-1.fc19.noarch and 
sos-3.0-23.fc19.noarch


I have used ovirt repo - yum localinstall 
http://resources.ovirt.org/releases/ovirt-release.noarch.rpm




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] FC19 - Transaction check error - while installing ovirt-engine

2014-04-23 Thread Thomas Suckow

Ok, I managed to update with:  yum --exclude=sos/* update


I am seeing this on upgrade as well. So far I have not found a solution.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] qemu-kvm-rhev for el6

2014-04-23 Thread Paul Jansen
hello,

does anyone know if there are an existent bugzilla to track the release 
of qemu-kvm-rhev rpms under el (like centos)?
Because I've looked at bugzilla and google docs oVirt Planning  
Tracking with no luck

best regards
a
I think this is the best fit that I have found so far: 
https://bugzilla.redhat.com/show_bug.cgi?id=1009100
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] configuration maximus

2014-04-23 Thread Itamar Heim

On 04/23/2014 11:49 PM, Ovirt User wrote:

Hello,

1) how many Vm per node are supported ?


should be ok with a couple of hundreds.


2) how many virtual CPU per node are supported ?


virtual cpu are per VM? should be ok with 100 per VM (assuming you have 
a machine with enough cores!).
I doubt anyone tested maxing these two figures out together, i don't see 
a relevant use case.





i don't find configuration maximus ..

thanks
Lukas
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Spurious error message when the RHEV capacity is maxed out?

2014-04-23 Thread Itamar Heim

On 04/24/2014 12:31 AM, Satya Vempati wrote:

We have about 900 vms in a RHEV-M 3.3 installation (3.3.2-0.50.el6ev)
with three hosts.

Each VM has 1 socket, 1 core for CPU and 256 MB memory and 2 x 1TB disks
+ one 5 GB boot disk.

Each host has 2 cpu sockets with 4 cpu cores per socket.

When we power on about 250 of these vms, we cannot power on any more vms.


is this a 'monday morning effect' (are you trying to launch them all 
together)?




Attempting to do so gets the error (in the RHEV-M) console.

Bad volume specification {'index': 0, 'iface': 'virtio', 'reqsize': '0',
'format': 'raw', 'bootOrder': '1', 'volumeID':
'14aa7302-7f6d-4b80-922d-3fc090695447', 'apparentsize': '5368709120',
'imageID': '97b1985e-e507-4927-91e9-7393cde5', 'specParams': {},
'readonly': 'false', 'domainID': '0c7abccc-1b97-4b39-ab96-e977a19fac3a',
'optional': 'false', 'deviceId': '97b1985e-e507-4927-91e9-7393cde5',
'truesize': '0', 'poolID': '5849b030-626e-47cb-ad90-3ce782d831b3',
'device': 'disk', 'shared': 'false', 'propagateErrors': 'off', 'type':
'disk'}.

This looks like a spurious error, as we can power on the same VM after
rebooting the hosts.

And also, the hosts do not really look maxed out (CPU usage is around
25%, memory around 50%), but no more vm’s can be powered on.

I was looking for a sizing guide, and found only
https://access.redhat.com/site/sites/default/files/attachments/rhev_sizing_0812.pdf

Is there any other information available?



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] FC19 - Transaction check error - while installing ovirt-engine

2014-04-23 Thread Itamar Heim

On 04/24/2014 03:21 AM, Thomas Suckow wrote:

Ok, I managed to update with:  yum --exclude=sos/* update


I am seeing this on upgrade as well. So far I have not found a solution.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


is there a bug tracking this?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Foreman not starting VM, hangs on finishing creating a new VM

2014-04-23 Thread Itamar Heim

On 04/23/2014 09:50 PM, Matt . wrote:

Hi All,

I'm having an issue with oVirt 3.4 and Foreman 1.5 RC1 at the moment, it
does not start/finish a new provisioned VM/Host., it creates it well on
oVirt.

I'm sure this worked well on a nightly some dags ago and I didn't update
it after it.

The foreman bug is shown here: http://projects.theforeman.org/issues/5132

What I also see in the engine log is a bunch of admin login errors which
I cannot place anywhere:

2014-04-23 20:26:44,203 INFO
[org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-6)
Running command: LoginUserCommand internal: false.
2014-04-23 20:26:44,242 INFO
[org.ovirt.engine.core.bll.LogoutUserCommand] (ajp--127.0.0.1-8702-6)
[2e4a8183] Running command: LogoutUserCommand internal: false.
2014-04-23 20:26:44,276 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-6) [2e4a8183] Correlation ID: 2e4a8183, Call Stack:
null, Custom Event ID: -1, Message: User admin logged out.
2014-04-23 20:26:44,470 INFO
[org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-1)
Running command: LoginUserCommand internal: false.
2014-04-23 20:26:44,510 INFO
[org.ovirt.engine.core.bll.LogoutUserCommand] (ajp--127.0.0.1-8702-1)
[5d731a43] Running command: LogoutUserCommand internal: false.
2014-04-23 20:26:44,542 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-1) [5d731a43] Correlation ID: 5d731a43, Call Stack:
null, Custom Event ID: -1, Message: User admin logged out.
2014-04-23 20:26:44,575 INFO
[org.ovirt.engine.core.bll.LoginUserCommand] (ajp--127.0.0.1-8702-3)
Running command: LoginUserCommand internal: false.
2014-04-23 20:26:44,623 INFO
[org.ovirt.engine.core.bll.RemoveVmCommand] (ajp--127.0.0.1-8702-3)
[708f0b9a] Lock Acquired to object EngineLock [exclusiveLocks= key:
7bfc175c-dbcd-43c9-9549-cde8d3b3b731 value: VM

I get the feeling this has something todo with eachother.



why do you think above snippet contain an error?

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] does SPM can run over ovirt-engine host ?

2014-04-23 Thread Itamar Heim

On 04/23/2014 09:17 PM, Tamer Lima wrote:

Hi , Yair

yes, I dont want to re-install ovirt

I'm not sure if allInone could be fix this problem.

allinone install vdsm on the same ovirt-engine host. Well, I already
have this : ovirt-engine, vdsm on the same host: srv-0202

My storage domains(data and iso) are hosted on srv-0202


I believe my solution is create one engine per server. Three independent
engines managing only local virtual machines.


this does not sounds right.
engine and SPM should not communicate at 98% traffic for 2 hours.
SPM should be one of the nodes in the DC.
engine isn't acting as a node (even if deployed on one by chance of 
all-in-one or hosted engine).


are you creating the VMs from template thinly provisioned or clone?








On Mon, Apr 14, 2014 at 10:07 PM, Yair Zaslavsky yzasl...@redhat.com
mailto:yzasl...@redhat.com wrote:

Hi Tamer,
Are you familiar with the all in one feature?

http://www.ovirt.org/Feature/AllInOne

I'm not sure if this can help you now, as you probably  don't want
to re-install ovirt, right?


- Original Message -
  From: Tamer Lima tamer.amer...@gmail.com
mailto:tamer.amer...@gmail.com
  To: users@ovirt.org mailto:users@ovirt.org
  Sent: Monday, April 14, 2014 5:13:12 PM
  Subject: [ovirt-users] does SPM can run over ovirt-engine host ?
 
  Hello,
 
  When I create virtual machine from a template (centos6.5, 2
cores, 8GB mem,
  500GB hd)  this process takes almost 2 hours.   I click on New
VM button
  and just select the template and click ok.
 
  engine.log show me high network consumption (98%)  between
  engine-server
  host and SPM host.
 
  I tried to make my engine-server host a spm host too, but without
sucess.
 
 
  Does SPM can run over on the same ovirt-engine machine ?
 
  Am I make something wrong? Or create VM from template is really
slow ?
 
 
  my servers :
  srv-0202  = ovirt-engine  , vdsm
  srv-0203 = spm , vdsm
  srv-0204 = vdsm
  These servers are dell blades connected on a 100GB switch.
 
 
 
  thanks
 
 
 
 
  This is what I know about SPM:
  http://www.ovirt.org/Storage_-_oVirt_workshop_November_2011
 
  = Storage Pool Manager (SPM) A role assigned to one host in a
data center
  granting it sole authority over:
 
 - Creation, deletion, an dmanipulation of virtula disk images,
snapshots
 and templates
- Templates: you can create on VM as a golden image and
provision to
multiple VMs (QCOW layers)
 - Allocation of storage for sparse block devices (on SAN)
- Thin provisinoing (see below)
 - Single metadata writer:
- SPM lease mechanism (Chockler and Malkhi 2004,
Light-Weight Leases
for Storage-Cnntric Coordination)
- Storage-centric mailbox
 - This role can be migrated to any host in data center
 
  ___
  Users mailing list
  Users@ovirt.org mailto:Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
 




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users