Re: [ovirt-users] Communication errors between engine and nodes?

2015-03-19 Thread Gabi C
Hello!

FYI:
 updated 2 days ago 3 hypervisor from my setup to latest 3.5-patternfly ,
rebooted nodes, and engine and the error seems to be gone: no longer got
heartbeat exeeded.

On Tue, Mar 17, 2015 at 11:58 AM, Piotr Kliczewski 
piotr.kliczew...@gmail.com wrote:

 Hi Roel,

 You can change this setting in two ways.
 - you can update it in db directly as you stated (not recommended)
 - use engine-config -s vdsHeartbeatInSeconds=20 but prior to running
 this command
   you need to update config file
 /etc/ovirt-engine/engine-config/engine-config.properties
   with vdsHeartbeatInSeconds.type=Integer. This config value is not
 exposed by default.

 Thanks,
 Piotr

 On Mon, Mar 16, 2015 at 11:18 PM, Roel de Rooy rder...@motto.nl wrote:
  HI Piotr,
 
  Thanks for your reply!
 
  If I would like to change the heartbeat value, do I have to update the
 value within the vdc_options table directly, or should this be done by
 another way (e.g. config file)?
 
  Regards,
  Roel
 
  -Oorspronkelijk bericht-
  Van: Piotr Kliczewski [mailto:piotr.kliczew...@gmail.com]
  Verzonden: maandag 16 maart 2015 12:16
  Aan: Roel de Rooy
  CC: Michal Skrivanek; users@ovirt.org
  Onderwerp: Re: [ovirt-users] Communication errors between engine and
 nodes?
 
  Unfortunately log entries that you copied give me almost no information
 about nature of your issue.
  There are few things that we can do to understand what is going on with
 your setup.
 
  Heartbeat functionality provides means to detect whether we still have
 connection with a host. By default heartbeat timeout is set to 10 seconds
 but it can be modified by setting vdsHeartbeatInSeconds.
 
  In general whenever there are no incoming responses nor heartbeat frame
 is not received engine will invalidate the connection and will attempt to
 recover. If reconnection was successful you want see any other consequences
 of loosing single heartbeat. I would explore stability of your network so
 if the network is busy or you loose network packets from time to time this
 kind of entries in the log are expected. You can increase heatbeat value
 and see whether it will work better for your env.
 
  If you confirm that your network is stable we could explore the issue
 further by setting debug level logging for your engine to understand
 exactly how the messages are processes by a host and when we receive
 responses.
 
 
 
  On Mon, Mar 16, 2015 at 11:34 AM, Roel de Rooy rder...@motto.nl wrote:
  Received the heartbeat exeeded continuously this morning (seems to be
 quiet again for now).
  VM's still continue to work correctly and the storage domains (NFS
 shares) are still connected and reachable on the nodes, at the exact time
 that this issue is happening.
 
  Contacted our network engineer to see if he could see a load increase
 on our network, or could see any latency, errors, etc.
  Unfortunately he could not detect anything yet (he is still
 investigating this).
 
 
  I have attached both the engine and vdsm logs
 
  Engine.log:
 
  2015-03-16 10:10:10,506 ERROR
  [org.ovirt.engine.core.vdsbroker.vdsbroker.ListVDSCommand]
  (DefaultQuartzScheduler_Worker-45) [6d40f562] Command
  ListVDSCommand(HostName = HOST, HostId =
  3b87597e-081b-4c89-9b1e-cb04203259f5,
  vds=Host[HOST,3b87597e-081b-4c89-9b1e-cb04203259f5]) execution
  failed. Exception: VDSNetworkException: VDSGenericException:
  VDSNetworkException: Heartbeat exeeded
  2015-03-16 10:10:10,507 ERROR
  [org.ovirt.engine.core.vdsbroker.vdsbroker.SpmStatusVDSCommand]
  (DefaultQuartzScheduler_Worker-35) [2c53103c] Command
  SpmStatusVDSCommand(HostName = HOST, HostId =
  3b87597e-081b-4c89-9b1e-cb04203259f5, storagePoolId =
  124ae76f-8acb-412e-91cc-dff9f6ec665d) execution failed. Exception:
  VDSNetworkException: VDSGenericException: VDSNetworkException:
  Heartbeat exeeded
  2015-03-16 10:10:10,506 WARN
  [org.ovirt.vdsm.jsonrpc.client.internal.ResponseWorker]
  (ResponseWorker) Exception thrown during message processing
  2015-03-16 10:10:10,507 WARN
 [org.ovirt.engine.core.vdsbroker.VdsManager]
 (DefaultQuartzScheduler_Worker-45) [6d40f562] Host HOST is not
 responding. It will stay in Connecting state for a grace period of 88
 seconds and after that an attempt to fence the host will be issued.
  2015-03-16 10:10:10,510 INFO
  [org.ovirt.engine.core.bll.storage.SetStoragePoolStatusCommand]
  (DefaultQuartzScheduler_Worker-35) [7e61eee] Running command:
  SetStoragePoolStatusCommand internal: true. Entities affected :  ID:
  124ae76f-8acb-412e-91cc-dff9f6ec665d Type: StoragePool
  2015-03-16 10:10:10,512 INFO
  [org.ovirt.engine.core.vdsbroker.storage.StoragePoolDomainHelper]
  (DefaultQuartzScheduler_Worker-35) [7e61eee] Storage Pool
  124ae76f-8acb-412e-91cc-dff9f6ec665d - Updating Storage Domain
  bfa86142-6f2e-44fe-8a9c-cf4390f3b8ae status from Active to Unknown,
  reason : null
  2015-03-16 10:10:10,513 INFO
  [org.ovirt.engine.core.vdsbroker.storage.StoragePoolDomainHelper]
  

Re: [ovirt-users] VM failed to start | Bad volume specification

2015-03-19 Thread Michal Skrivanek

On Mar 19, 2015, at 03:18 , Punit Dambiwal hypu...@gmail.com wrote:

 Hi All,
 
 Is there any one have any idea about this problem...it seems it's bug either 
 in Ovirt or Glusterfs...that's why no one has the idea about itplease 
 correct me if i am wrong….

Hi,
as I said, storage access times out; so it seems to me as a gluster setup 
problem, the storage domain you have your VMs on is not working…

Thanks,
michal

 
 Thanks,
 Punit
 
 On Wed, Mar 18, 2015 at 5:05 PM, Punit Dambiwal hypu...@gmail.com wrote:
 Hi Michal,
 
 Would you mind to let me know the possible messedup things...i will check and 
 try to resolve itstill i am communicating gluster community to resolve 
 this issue...
 
 But in the ovirtgluster setup is quite straightso how come it will be 
 messedup with reboot ?? if it can be messedup with reboot then it seems not 
 good and stable technology for the production storage
 
 Thanks,
 Punit
 
 On Wed, Mar 18, 2015 at 3:51 PM, Michal Skrivanek 
 michal.skriva...@redhat.com wrote:
 
 On Mar 18, 2015, at 03:33 , Punit Dambiwal hypu...@gmail.com wrote:
 
  Hi,
 
  Is there any one from community can help me to solve this issue...??
 
  Thanks,
  Punit
 
  On Tue, Mar 17, 2015 at 12:52 PM, Punit Dambiwal hypu...@gmail.com wrote:
  Hi,
 
  I am facing one strange issue with ovirt/glusterfsstill didn't find 
  this issue is related with glusterfs or Ovirt
 
  Ovirt :- 3.5.1
  Glusterfs :- 3.6.1
  Host :- 4 Hosts (Compute+ Storage)...each server has 24 bricks
  Guest VM :- more then 100
 
  Issue :- When i deploy this cluster first time..it work well for me(all the 
  guest VM created and running successfully)but suddenly one day my one 
  of the host node rebooted and none of the VM can boot up now...and failed 
  with the following error Bad Volume Specification
 
  VMId :- d877313c18d9783ca09b62acf5588048
 
  VDSM Logs :- http://ur1.ca/jxabi
 
 you've got timeouts while accessing storage…so I guess something got messed 
 up on reboot, it may also be just a gluster misconfiguration…
 
  Engine Logs :- http://ur1.ca/jxabv
 
  
  [root@cpu01 ~]# vdsClient -s 0 getVolumeInfo 
  e732a82f-bae9-4368-8b98-dedc1c3814de 0002-0002-0002-0002-0145 
  6d123509-6867-45cf-83a2-6d679b77d3c5 9030bb43-6bc9-462f-a1b9-f6d5a02fb180
  status = OK
  domain = e732a82f-bae9-4368-8b98-dedc1c3814de
  capacity = 21474836480
  voltype = LEAF
  description =
  parent = ----
  format = RAW
  image = 6d123509-6867-45cf-83a2-6d679b77d3c5
  uuid = 9030bb43-6bc9-462f-a1b9-f6d5a02fb180
  disktype = 2
  legality = LEGAL
  mtime = 0
  apparentsize = 21474836480
  truesize = 4562972672
  type = SPARSE
  children = []
  pool =
  ctime = 1422676305
  -
 
  I opened same thread earlier but didn't get any perfect answers to solve 
  this issue..so i reopen it...
 
  https://www.mail-archive.com/users@ovirt.org/msg25011.html
 
  Thanks,
  Punit
 
 
 
 
 
 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Windows Server 2012 R2 guest + Memory Ballooning = High CPU

2015-03-19 Thread Vinzenz Feenstra

On 03/12/2015 05:34 PM, Doron Fediuck wrote:


On 11/03/15 12:38, Winfried de Heiden wrote:

Hi All,

Running Windows 2012 R2 (64 bit) on oVirt 3.5.1-1.el6 caused a rather
high cpu load, even ~20% when doing nothing. Turning off theMemory
Balloon Device made the problem disappear.

The guest tools are installed using the ovirt-guest-tools-3.5_5.iso.

Anyone got the same issue? What is causing this? Seems like a bug to me.
Yes it does indeed seem to be a bug, and from how this sounds like, I 
would suspect the windows ballooning driver.


Kind regards,

Winny



Hi Winny,
Can you please open a bug with the relevant versions you have?

Thanks,
Doron
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



--
Regards,

Vinzenz Feenstra | Senior Software Engineer
RedHat Engineering Virtualization R  D
Phone: +420 532 294 625
IRC: vfeenstr or evilissimo

Better technology. Faster innovation. Powered by community collaboration.
See how it works at redhat.com

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Power Management config on Ovirt

2015-03-19 Thread Renchu Mathew
Hi Eli/Martin,

On each host power management I have given the other hosts iRMC IPMI address. 
Is that correct or I have to give same hosts management IP? Ovirt engine is 
installed on another machine and it can reach the iRMC management IP of each 
host. My setup is as below.

2 hosts with 6 NICs configured as bond0 on each host. Another IP given for iRMC 
management port on each host and this IP is configured on the other host power 
management. VM stored on glusterfs storage and engine installed on another 
server. I can able to create VM and manually migrate but if one host is down, 
VM is not migrating to other host. What do I add more to achieve this.

Regards

Renchu Mathew  

-Original Message-
From: Eli Mesika [mailto:emes...@redhat.com] 
Sent: Wednesday, March 18, 2015 5:51 PM
To: Renchu Mathew
Cc: Martin Perina; users@ovirt.org
Subject: Re: [ovirt-users] Power Management config on Ovirt



- Original Message -
 From: Renchu Mathew ren...@cracknell.com
 To: Eli Mesika emes...@redhat.com
 Cc: Martin Perina mper...@redhat.com, users@ovirt.org
 Sent: Wednesday, March 18, 2015 3:15:40 PM
 Subject: RE: [ovirt-users] Power Management config on Ovirt
 
 Hi Eli,
 
 Those 2 hosts are connected with Fujitsu iRMC management port and power
 management is configured with ipmi. So it can use this connection to fence
 the other node, is it correct?

No, keep in mind that the one that communicates with the proxy host is the 
oVirt engine, so , if it is not accessable, oVirt engine can not use it 


 
 Regards
 
 Renchu Mathew
 
 -Original Message-
 From: Eli Mesika [mailto:emes...@redhat.com]
 Sent: Wednesday, March 18, 2015 4:31 PM
 To: Renchu Mathew
 Cc: Martin Perina; users@ovirt.org
 Subject: Re: [ovirt-users] Power Management config on Ovirt
 
 
 
 - Original Message -
  From: Renchu Mathew ren...@cracknell.com
  To: Martin Perina mper...@redhat.com
  Cc: users@ovirt.org
  Sent: Wednesday, March 18, 2015 2:24:46 PM
  Subject: Re: [ovirt-users] Power Management config on Ovirt
  
  Hi Martin,
  
  My setup meets all those requirements and I can able to migrate the VM
  from one host to another manually. Once network cable is pulled off
  from one of the server other server is also shuts down.
 
 Hi
 Sorry for jumping in late , Yesterday was an election day in ISRAEL...
 
 If the other server shuts down when you plug-off the first one and you have
 only 2 hosts then no fencing will take place since there is no available
 proxy host to perform the operation
 
  
  Regards
  
  Renchu Mathew  |  Sr. IT Administrator
  
  
  
  CRACKNELL  DUBAI   |  P.O. Box 66231  |   United Arab Emirates  |  T
  +971 4
  3445417  |  F +971 4 3493675 |  M +971 50 7386484 ABU DHABI | DUBAI |
  LONDON | MUSCAT | DOHA | JEDDAH EMAIL ren...@cracknell.com | WEB
  www.cracknell.com
  
  This email, its content and any files transmitted with it are intended
  solely for the addressee(s) and may be legally privileged and/or
  confidential. If you are not the intended recipient please let us know
  by email reply and delete it from the system. Please note that any
  views or opinions presented in this email do not necessarily represent
  those of the company. Email transmissions cannot be guaranteed to be
  secure or error-free as information could be intercepted, corrupted,
  lost, destroyed, arrive late or incomplete, or contain viruses. The
  company therefore does not accept liability for any errors or
  omissions in the contents of this message which arise as a result of email
  transmission.
  
  
  -Original Message-
  From: Martin Perina [mailto:mper...@redhat.com]
  Sent: Tuesday, March 17, 2015 8:31 PM
  To: Renchu Mathew
  Cc: users@ovirt.org
  Subject: Re: [ovirt-users] Power Management config on Ovirt
  
  Hi,
  
  prior to the test I would check this:
  
- Data Center status is Up
- All hosts status is Up
- All storage domains status is Up
- VM is running
  
  If this is valid, you can start your fence testing. But bear in mind
  what I sent you in previous email: at least one host in DC should be
  fully functional to be able to fence non responsive host.
  
  Martin Perina
  
  - Original Message -
   From: Renchu Mathew ren...@cracknell.com
   To: Martin Perina mper...@redhat.com
   Cc: users@ovirt.org
   Sent: Tuesday, March 17, 2015 5:03:53 PM
   Subject: RE: [ovirt-users] Power Management config on Ovirt
   
   Hi Martin
   
   Yes, my test VM still running on this storage. Is it possible to do
   remote session and check this?
   
   Regards
   
   Renchu Mathew
   
   
   -Original Message-
   From: Martin Perina [mailto:mper...@redhat.com]
   Sent: Tuesday, March 17, 2015 7:30 PM
   To: Renchu Mathew
   Cc: users@ovirt.org
   Subject: Re: [ovirt-users] Power Management config on Ovirt
   
   Hi,
   
   this is what happened (at least what I was able to read from log):
   
   18:18:02 - host node02 changed status to Connecting
   
 - 

Re: [ovirt-users] VM failed to start | Bad volume specification

2015-03-19 Thread Punit Dambiwal
Hi Joop,

No split brainbut i have find some disconnecting error in the brick
logs :-

---
[2015-03-19 09:47:38.201769] I [login.c:82:gf_auth] 0-auth/login: allowed
user names: 6ba6d773-cf70-4ba9-98d3-d2e31bcbb14c
[2015-03-19 09:47:38.201819] I [server-handshake.c:585:server_setvolume]
0-ds01-server: accepted client from
cpu01-13039-2015/03/19-09:47:37:502523-ds01-client-0-0-0 (version: 3.6.1)
[2015-03-19 09:47:38.243512] I [server.c:518:server_rpc_notify]
0-ds01-server: disconnecting connection from
cpu01-13039-2015/03/19-09:47:37:502523-ds01-client-0-0-0
[2015-03-19 09:47:38.243579] I [client_t.c:417:gf_client_unref]
0-ds01-server: Shutting down connection
cpu01-13039-2015/03/19-09:47:37:502523-ds01-client-0-0-0
-

On Thu, Mar 19, 2015 at 5:42 PM, Joop jvdw...@xs4all.nl wrote:

 Please keep it on the list.

 On 19-3-2015 10:10, Punit Dambiwal wrote:
  Hi Joop,
 
  How i can check the split-brain or not ???i already reblance the
  storage nodes and also run the heal command for the consistency
 
 gluster volume heal YourVolume info split-brain

 Regards,

 Joop


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] VM failed to start | Bad volume specification

2015-03-19 Thread Punit Dambiwal
Hi Michal,

The Storage domain is up and running and mounted on all the host nodes...as
i updated before that it was working perfectly before but just after reboot
can not make the VM poweron...

[image: Inline image 1]

[image: Inline image 2]

[root@cpu01 log]# gluster volume info

Volume Name: ds01
Type: Distributed-Replicate
Volume ID: 369d3fdc-c8eb-46b7-a33e-0a49f2451ff6
Status: Started
Number of Bricks: 48 x 2 = 96
Transport-type: tcp
Bricks:
Brick1: cpu01:/bricks/1/vol1
Brick2: cpu02:/bricks/1/vol1
Brick3: cpu03:/bricks/1/vol1
Brick4: cpu04:/bricks/1/vol1
Brick5: cpu01:/bricks/2/vol1
Brick6: cpu02:/bricks/2/vol1
Brick7: cpu03:/bricks/2/vol1
Brick8: cpu04:/bricks/2/vol1
Brick9: cpu01:/bricks/3/vol1
Brick10: cpu02:/bricks/3/vol1
Brick11: cpu03:/bricks/3/vol1
Brick12: cpu04:/bricks/3/vol1
Brick13: cpu01:/bricks/4/vol1
Brick14: cpu02:/bricks/4/vol1
Brick15: cpu03:/bricks/4/vol1
Brick16: cpu04:/bricks/4/vol1
Brick17: cpu01:/bricks/5/vol1
Brick18: cpu02:/bricks/5/vol1
Brick19: cpu03:/bricks/5/vol1
Brick20: cpu04:/bricks/5/vol1
Brick21: cpu01:/bricks/6/vol1
Brick22: cpu02:/bricks/6/vol1
Brick23: cpu03:/bricks/6/vol1
Brick24: cpu04:/bricks/6/vol1
Brick25: cpu01:/bricks/7/vol1
Brick26: cpu02:/bricks/7/vol1
Brick27: cpu03:/bricks/7/vol1
Brick28: cpu04:/bricks/7/vol1
Brick29: cpu01:/bricks/8/vol1
Brick30: cpu02:/bricks/8/vol1
Brick31: cpu03:/bricks/8/vol1
Brick32: cpu04:/bricks/8/vol1
Brick33: cpu01:/bricks/9/vol1
Brick34: cpu02:/bricks/9/vol1
Brick35: cpu03:/bricks/9/vol1
Brick36: cpu04:/bricks/9/vol1
Brick37: cpu01:/bricks/10/vol1
Brick38: cpu02:/bricks/10/vol1
Brick39: cpu03:/bricks/10/vol1
Brick40: cpu04:/bricks/10/vol1
Brick41: cpu01:/bricks/11/vol1
Brick42: cpu02:/bricks/11/vol1
Brick43: cpu03:/bricks/11/vol1
Brick44: cpu04:/bricks/11/vol1
Brick45: cpu01:/bricks/12/vol1
Brick46: cpu02:/bricks/12/vol1
Brick47: cpu03:/bricks/12/vol1
Brick48: cpu04:/bricks/12/vol1
Brick49: cpu01:/bricks/13/vol1
Brick50: cpu02:/bricks/13/vol1
Brick51: cpu03:/bricks/13/vol1
Brick52: cpu04:/bricks/13/vol1
Brick53: cpu01:/bricks/14/vol1
Brick54: cpu02:/bricks/14/vol1
Brick55: cpu03:/bricks/14/vol1
Brick56: cpu04:/bricks/14/vol1
Brick57: cpu01:/bricks/15/vol1
Brick58: cpu02:/bricks/15/vol1
Brick59: cpu03:/bricks/15/vol1
Brick60: cpu04:/bricks/15/vol1
Brick61: cpu01:/bricks/16/vol1
Brick62: cpu02:/bricks/16/vol1
Brick63: cpu03:/bricks/16/vol1
Brick64: cpu04:/bricks/16/vol1
Brick65: cpu01:/bricks/17/vol1
Brick66: cpu02:/bricks/17/vol1
Brick67: cpu03:/bricks/17/vol1
Brick68: cpu04:/bricks/17/vol1
Brick69: cpu01:/bricks/18/vol1
Brick70: cpu02:/bricks/18/vol1
Brick71: cpu03:/bricks/18/vol1
Brick72: cpu04:/bricks/18/vol1
Brick73: cpu01:/bricks/19/vol1
Brick74: cpu02:/bricks/19/vol1
Brick75: cpu03:/bricks/19/vol1
Brick76: cpu04:/bricks/19/vol1
Brick77: cpu01:/bricks/20/vol1
Brick78: cpu02:/bricks/20/vol1
Brick79: cpu03:/bricks/20/vol1
Brick80: cpu04:/bricks/20/vol1
Brick81: cpu01:/bricks/21/vol1
Brick82: cpu02:/bricks/21/vol1
Brick83: cpu03:/bricks/21/vol1
Brick84: cpu04:/bricks/21/vol1
Brick85: cpu01:/bricks/22/vol1
Brick86: cpu02:/bricks/22/vol1
Brick87: cpu03:/bricks/22/vol1
Brick88: cpu04:/bricks/22/vol1
Brick89: cpu01:/bricks/23/vol1
Brick90: cpu02:/bricks/23/vol1
Brick91: cpu03:/bricks/23/vol1
Brick92: cpu04:/bricks/23/vol1
Brick93: cpu01:/bricks/24/vol1
Brick94: cpu02:/bricks/24/vol1
Brick95: cpu03:/bricks/24/vol1
Brick96: cpu04:/bricks/24/vol1
Options Reconfigured:
diagnostics.count-fop-hits: on
diagnostics.latency-measurement: on
nfs.disable: on
user.cifs: enable
auth.allow: 10.10.0.*
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
storage.owner-uid: 36
storage.owner-gid: 36
server.allow-insecure: on
network.ping-timeout: 100
[root@cpu01 log]#

-

[root@cpu01 log]# gluster volume status
Status of volume: ds01
Gluster process PortOnline  Pid
--
Brick cpu01:/bricks/1/vol1  49152   Y
33474
Brick cpu02:/bricks/1/vol1  49152   Y
40717
Brick cpu03:/bricks/1/vol1  49152   Y
18080
Brick cpu04:/bricks/1/vol1  49152   Y
40447
Brick cpu01:/bricks/2/vol1  49153   Y
33481
Brick cpu02:/bricks/2/vol1  49153   Y
40724
Brick cpu03:/bricks/2/vol1  49153   Y
18086
Brick cpu04:/bricks/2/vol1  49153   Y
40453
Brick cpu01:/bricks/3/vol1  49154   Y
33489
Brick cpu02:/bricks/3/vol1  49154   Y
40731
Brick cpu03:/bricks/3/vol1  49154   Y
18097
Brick cpu04:/bricks/3/vol1  49154   Y
40460

Re: [ovirt-users] virtio-blk-data-plane

2015-03-19 Thread Michal Skrivanek

On Oct 27, 2014, at 09:09 , Wolfgang Bucher wolfgang.buc...@netland-mn.de 
wrote:

 Hello,
 is it possible to use virtio-blk-data-plane in ovirt with el7 host.

Hi,
no, not yet, planned for 3.6 though.

Thanks,
michal

 
 Thanks
 Wolfgang 
 
 
 
 
 
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] VM failed to start | Bad volume specification

2015-03-19 Thread Punit Dambiwal
Hi Joop,

How i can check the split-brain or not ???i already reblance the
storage nodes and also run the heal command for the consistency

On Thu, Mar 19, 2015 at 4:36 PM, Joop jvdw...@xs4all.nl wrote:

 On 19-3-2015 9:25, Punit Dambiwal wrote:
  Hi Michal,
 
  The Storage domain is up and running and mounted on all the host
  nodes...as i updated before that it was working perfectly before but
  just after reboot can not make the VM poweron...
 
 
 Did you check if you're suffering from a split-brain situation?
 My test setup sometimes does when not being careful when updating
 glusterfs/ovirt.

 Regards,

 Joop

 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] VM failed to start | Bad volume specification

2015-03-19 Thread Joop
Please keep it on the list.

On 19-3-2015 10:10, Punit Dambiwal wrote:
 Hi Joop,

 How i can check the split-brain or not ???i already reblance the
 storage nodes and also run the heal command for the consistency

gluster volume heal YourVolume info split-brain

Regards,

Joop


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] VM failed to start | Bad volume specification

2015-03-19 Thread Joop
On 19-3-2015 9:25, Punit Dambiwal wrote:
 Hi Michal,

 The Storage domain is up and running and mounted on all the host
 nodes...as i updated before that it was working perfectly before but
 just after reboot can not make the VM poweron...


Did you check if you're suffering from a split-brain situation?
My test setup sometimes does when not being careful when updating
glusterfs/ovirt.

Regards,

Joop

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Power Management config on Ovirt

2015-03-19 Thread Martin Perina
Hi,

- Original Message -
 From: Renchu Mathew ren...@cracknell.com
 To: Eli Mesika emes...@redhat.com
 Cc: Martin Perina mper...@redhat.com, users@ovirt.org
 Sent: Thursday, March 19, 2015 9:30:30 AM
 Subject: RE: [ovirt-users] Power Management config on Ovirt
 
 Hi Eli/Martin,
 
 On each host power management I have given the other hosts iRMC IPMI address.
 Is that correct or I have to give same hosts management IP? 

Each host should have its own IPMI address assigned in PM settings.

For example you have these two hosts:

  1. host1 with IP1 and IPMI1
  2. host2 with IP2 and IPMI2

So when you define a host in oVirt you should use these values in host
configuration:

  1. host1
   IP Address: IP1
   Power Management IP address: IPMI1

  2. host2
   IP Address: IP2
   Power Management IP address: IPMI2

 Ovirt engine is
 installed on another machine and it can reach the iRMC management IP of each
 host. My setup is as below.

This is correct, but it's not necessary for engine machine to access power
management IPs. But it's required that each host in the cluster have to
be able to access all other host power management interfaces. Engine will
select one functional host (fence proxy) and on this host power management
action is executed. For example you may have this scenario:

1. host1 is Up, connection to host2 is broken
2. engine will select host1 as fence proxy
3. On host1 power management commands are executed (status, stop, start)


 
 2 hosts with 6 NICs configured as bond0 on each host. Another IP given for
 iRMC management port on each host and this IP is configured on the other
 host power management. VM stored on glusterfs storage and engine installed
 on another server. I can able to create VM and manually migrate but if one
 host is down, VM is not migrating to other host. What do I add more to
 achieve this.

It looks OK except for power management IP address, see above.

Martin


 
 Regards
 
 Renchu Mathew
 
 -Original Message-
 From: Eli Mesika [mailto:emes...@redhat.com]
 Sent: Wednesday, March 18, 2015 5:51 PM
 To: Renchu Mathew
 Cc: Martin Perina; users@ovirt.org
 Subject: Re: [ovirt-users] Power Management config on Ovirt
 
 
 
 - Original Message -
  From: Renchu Mathew ren...@cracknell.com
  To: Eli Mesika emes...@redhat.com
  Cc: Martin Perina mper...@redhat.com, users@ovirt.org
  Sent: Wednesday, March 18, 2015 3:15:40 PM
  Subject: RE: [ovirt-users] Power Management config on Ovirt
  
  Hi Eli,
  
  Those 2 hosts are connected with Fujitsu iRMC management port and power
  management is configured with ipmi. So it can use this connection to fence
  the other node, is it correct?
 
 No, keep in mind that the one that communicates with the proxy host is the
 oVirt engine, so , if it is not accessable, oVirt engine can not use it
 
 
  
  Regards
  
  Renchu Mathew
  
  -Original Message-
  From: Eli Mesika [mailto:emes...@redhat.com]
  Sent: Wednesday, March 18, 2015 4:31 PM
  To: Renchu Mathew
  Cc: Martin Perina; users@ovirt.org
  Subject: Re: [ovirt-users] Power Management config on Ovirt
  
  
  
  - Original Message -
   From: Renchu Mathew ren...@cracknell.com
   To: Martin Perina mper...@redhat.com
   Cc: users@ovirt.org
   Sent: Wednesday, March 18, 2015 2:24:46 PM
   Subject: Re: [ovirt-users] Power Management config on Ovirt
   
   Hi Martin,
   
   My setup meets all those requirements and I can able to migrate the VM
   from one host to another manually. Once network cable is pulled off
   from one of the server other server is also shuts down.
  
  Hi
  Sorry for jumping in late , Yesterday was an election day in ISRAEL...
  
  If the other server shuts down when you plug-off the first one and you have
  only 2 hosts then no fencing will take place since there is no available
  proxy host to perform the operation
  
   
   Regards
   
   Renchu Mathew  |  Sr. IT Administrator
   
   
   
   CRACKNELL  DUBAI   |  P.O. Box 66231  |   United Arab Emirates  |  T
   +971 4
   3445417  |  F +971 4 3493675 |  M +971 50 7386484 ABU DHABI | DUBAI |
   LONDON | MUSCAT | DOHA | JEDDAH EMAIL ren...@cracknell.com | WEB
   www.cracknell.com
   
   This email, its content and any files transmitted with it are intended
   solely for the addressee(s) and may be legally privileged and/or
   confidential. If you are not the intended recipient please let us know
   by email reply and delete it from the system. Please note that any
   views or opinions presented in this email do not necessarily represent
   those of the company. Email transmissions cannot be guaranteed to be
   secure or error-free as information could be intercepted, corrupted,
   lost, destroyed, arrive late or incomplete, or contain viruses. The
   company therefore does not accept liability for any errors or
   omissions in the contents of this message which arise as a result of
   email
   transmission.
   
   
   -Original Message-
   From: Martin Perina 

Re: [ovirt-users] 答复: bonding 802.3ad mode

2015-03-19 Thread Dan Kenigsberg
On Wed, Mar 18, 2015 at 10:57:18PM -0400, Dan Yasny wrote:
 Mode 0 is not supported under a bridge, just like mode 6
 
 On Wed, Mar 18, 2015 at 10:47 PM, Xie, Chao xiec.f...@cn.fujitsu.com
 wrote:
 
   Yeah, Alex is right. And if you want to double the network’s speed in
  single flow, the mode 0 is only choice. But mode 0 seems not be supported
  in oVirt?
 
 
 
  *发件人:* users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] *代表 *Alex
  Crow
  *发送时间:* 2015年3月19日 0:25
  *收件人:* users@ovirt.org
  *主题:* Re: [ovirt-users] bonding 802.3ad mode
 
 
 
  The balancing on 802.3ad only occurs for different network flows based on
  a hash of source and destination MAC (or can be made to add IP addresses
  into the calculation). A single flow will only use a single NIC in ad mode.
 
  Alex
 
 
   On 18/03/15 16:17, Nathanaël Blanchet wrote:
 
  Hi all,
 
  I'm used to create a mode 4 bond0 interface with two 1 Gb/s interfaces on
  all my hosts, and ethtool bond0 gives me a functionnal 2000Mb/s. However,
  when importing a vm from the export domain (NFS with a speed of 4GB/s), I
  always have this alert:
 
  Host siple has network interface which exceeded the defined threshold
  [95%] (em3: transmit rate[0%], receive rate [100%])
  It seems that the second nic never works while the first one is overloaded.
  Is it an expected behaviour? I believed that the flow was balanced between
  the two interfaces in 802.3ad mode.

To follow up on former ressponses: what do you have on top of your bond?
If you have a VM network, multiple guests are expected to have a
different hash value for each, and to spread the load on mode 4.

If you use the bonds for a host network (e.g. dispaly, migration,
storage) you can try mode 0.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] test, please ignore

2015-03-19 Thread Mike
A post didn't reach the list, this is a test, please ignore. Mike


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] bonding 802.3ad mode

2015-03-19 Thread Lior Vernia
Bonjour Nathanael,

You haven't mentioned which version of oVirt you were using - I suspect
it's pre-3.5 and therefore this isn't fixed yet:
https://bugzilla.redhat.com/show_bug.cgi?id=1114085

From 3.5 onwards the warning should not appear for a bond (in
aggregating mode) if only one of its slaves is overloaded.

A bientot, Lior.

On 18/03/15 18:17, Nathanaël Blanchet wrote:
 Hi all,
 
 I'm used to create a mode 4 bond0 interface with two 1 Gb/s interfaces
 on all my hosts, and ethtool bond0 gives me a functionnal 2000Mb/s.
 However, when importing a vm from the export domain (NFS with a speed of
 4GB/s), I always have this alert:
 Host siple has network interface which exceeded the defined threshold
 [95%] (em3: transmit rate[0%], receive rate [100%])
 It seems that the second nic never works while the first one is overloaded.
 Is it an expected behaviour? I believed that the flow was balanced
 between the two interfaces in 802.3ad mode.
 
 
 
 
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users
 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Hosted Engine config in our hardware environment

2015-03-19 Thread Eric Wong
Hello oVirt guru out there:

I want to seek some advice on upgrade path for our oVirt management vm 
configuration.  We have been using oVirt for over 3 years.  When we first setup 
oVirt environment, Hosted Engine componment did not exist.  Our question is 
should we migrate our current configuration to use Hosted Engine?

First let me give an overview of our configuration.  We have blade servers in 2 
separate racks.  D1  X1.  Each side has 10 blades.  Storage is iSCSI SAN.

Inside our oVirt 3.5.0.1-1.el6 installation, it is configured with 2 data 
centers.  D1  X1.  Each datacenter has the 10 blades for that side.  The 
management function of oVirt (oVirt web console) is running off 2 VMs, 
ovirt-mgmt-1 on D1, and ovirt-mgmt-2 on X1.  We have keepalived to maintain a 
flowing IP for the oVirt management console.  The keepalived script makes sure 
only one copy of ovirt-engine is running at any time.  It can be on D1 or X1.  
The mgmt VMs have Postgresql setup in replication mode.  In case one of the 
mgmt vm failed, the other mgmt vm on the other rack can pick up the mgmt role.  
Both mgmt VMs can see all blades and SAN resources on D1  X1.

This configuration has been working well for us.  The drawback is if both ovirt 
mgmt vm crashed, we will not be able to start them or make any change to the 
ovirt environment.  It is because the mgmt VMs are running within the oVirt 
domain.

We tried to upgrade our configuration to Hosted Engine configuration.  From 
what I understand, the Hosted Engine will run in a separate storage domain.  In 
both times we tried to upgrade to Hosted Engine, they both failed during export 
and import of current configuration.

I think my questions are:
- will the Hosted Engine model works in our hardware configuration.  With 
hardware in 2 racks, D1  X1.  Can a single Hosted Engine manage hardware on 
both sides?
- How can we achieve redundancy when running Hosted Engine?  We need to have 
management capibility in case one side completely failed.

Thanks in advance.
Eric
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Cluster comms error, can't migrate

2015-03-19 Thread Mike
Hi Everyone

I have a two node hosted engine cluster that's been running for a month
or two. 

NFS is used for the VM's shared off the nodes on a second network
interface with different hostnames, I hope easier to migrate later on.
NFS 172.16.67.0/24 ov1-nfs.domain.dom on .1 and ov2-nfs.domain.dom
on .2. The NFS shares are working.

Management net is 10.10.10.224/28

Last night the cluster had communication errors, but I could not find
any issues, all nodes can ping  ssh with each other and engine.

Today, it got worse, the engine migrated all but 3 VM's to OV2, the node
hosting the engine. The VMs still on OV1 are there because the migration
for those failed. I can't manually migrate anything back to ov1. I
eventually shut down the engine and started on OV1, but still no joy.

The VMs are alive, both on OV1  OV2. OV2 is currently in local
maintenance to stop the engine moving and stop the email alerts.

I have been through the logs, I see there be a cert issue in
libvirtd.log on
the receiving host?

The web interface cert has been changed for a CA issued one several
weeks ago, this is working and I monitor the manager using nagios via
the API. This is still working. I did not change any of the other
internal oVirt PKI stuff.

The portal is working to view the VMs and config, but I haven't tried
changing anything during this issue apart from trying the migrate. The
VMs are providing public services, so I have not tried changing things.

Linux ov1.domain.dom 2.6.32-504.3.3.el6.x86_64 #1 SMP Wed Dec 17 01:55:02 UTC 
2014 x86_64 x86_64 x86_64 GNU/Linux
oVirt Engine Version: 3.5.1-1.el6
OS Version: RHEL - 6 - 6.el6.centos.12.2
Kernel Version: 2.6.32 - 504.3.3.el6.x86_64
KVM Version: 0.12.1.2 - 2.448.el6_6
LIBVIRT Version: libvirt-0.10.2-46.el6_6.2
VDSM Version: vdsm-4.16.10-8.gitc937927.el6
SPICE Version: 0.12.4 - 11.el6

Any help appreciated.
Mike


[root@ov1 ~]#  libvirtd.log
2015-03-18 15:42:17.387+: 3017: error :
virNetTLSContextValidCertificate:1008 : Unable to verify TLS peer: The
peer did not send any certificate.

2015-03-18 15:42:17.387+: 3017: warning :
virNetTLSContextCheckCertificate:1142 : Certificate check failed Unable
to verify TLS peer: The peer did not send any certificate.

2015-03-18 15:42:17.387+: 3017: error :
virNetTLSContextCheckCertificate:1145 : authentication failed: Failed to
verify peer's certificate

[root@ov2 ~]#  vdsm.log
Thread-49490::DEBUG::2015-03-18
15:42:17,294::migration::298::vm.Vm::(_startUnderlyingMigration)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::starting migration to qemu
+tls://ov1.domain.dom/system with miguri tcp://10.10.10.227

Thread-49525::DEBUG::2015-03-18
15:42:17,296::migration::361::vm.Vm::(run)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::migration downtime thread
started

Thread-49526::DEBUG::2015-03-18
15:42:17,297::migration::410::vm.Vm::(monitor_migration)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::starting migration monitor
thread

Thread-49490::DEBUG::2015-03-18
15:42:17,388::libvirtconnection::143::root::(wrapper) Unknown
libvirterror: ecode: 9 edom: 10 level: 2 message: operation failed:
Failed to connect to remote libvirt URI qemu+tls://ov1.domain.dom/system

Thread-49490::DEBUG::2015-03-18
15:42:17,390::migration::376::vm.Vm::(cancel)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::canceling migration
downtime thread

Thread-49525::DEBUG::2015-03-18
15:42:17,391::migration::373::vm.Vm::(run)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::migration downtime thread
exiting

Thread-49490::DEBUG::2015-03-18
15:42:17,391::migration::470::vm.Vm::(stop)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::stopping migration monitor
thread

Thread-49490::ERROR::2015-03-18
15:42:17,393::migration::161::vm.Vm::(_recover)
vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::operation failed: Failed to
connect to remote libvirt URI qemu+tls://ov1.domain.dom/system


[root@ov1 ~]# cat /var/log/vdsm/vdsm.log|grep MY_VM
Thread-7589263::DEBUG::2015-03-18
15:22:01,936::BindingXMLRPC::1133::vds::(wrapper) client
[10.10.10.228]::call vmMigrationCreate with ({'status': 'Up',
'acpiEnable': 'true', 'emulatedMachine': 'rhel6.5.0',
'afterMigrationStatus': '', 'tabletEnable': 'true', 'vmId':
'b44b2182-f943-4987-8421-8a98fd2a04d4', 'memGuaranteedSize': 2048,
'transparentHugePages': 'true', 'displayPort': '5929',
'displaySecurePort': '-1', 'spiceSslCipherSuite': 'DEFAULT', 'cpuType':
'SandyBridge', 'smp': '2', 'migrationDest': 'libvirt', 'custom': {},
'vmType': 'kvm', '_srcDomXML': domain type='kvm' id='58'\n
nameMY_VM/name\n
uuidb44b2182-f943-4987-8421-8a98fd2a04d4/uuid\n  memory
unit='KiB'2097152/memory\n  currentMemory
unit='KiB'2097152/currentMemory\n  vcpu placement='static'
current='2'16/vcpu\n  cputune\nshares1020/shares\n
/cputune\n  sysinfo type='smbios'\nsystem\n  entry
name='manufacturer'oVirt/entry\n  entry name='product'oVirt
Node
 /entry\n  entry name='version'6-6.el6.centos.12.2/entry\n
entry 

[ovirt-users] oVirt Newsletter Special Edition: Mid-March 2015

2015-03-19 Thread Brian Proffitt
It's a nice problem to have: it's been so busy in the oVirt community, we have 
to push out a mid-month newsletter just to keep up!


In the Community


Solving (NP-Hard) Scheduling Problems with oVirt and OptaPlanner 
http://www.socallinuxexpo.org/scale/13x/presentations/solving-np-hard-scheduling-problems-ovirt-and-optaplanner

FOSDEM Virtualization Room Videos:

Docker Integration in oVirt and IaaS 
https://video.fosdem.org/2015/devroom-virtualisation/dockerovirt__CAM_ONLY.mp4

Live migration for containers is around the corner 
https://video.fosdem.org/2015/devroom-virtualisation/livemigration__CAM_ONLY.mp4

LXD: The Container Hypervisor 
https://video.fosdem.org/2015/devroom-virtualisation/lxd.mp4

Observability in KVM--How to troubleshoot virtual machines 
https://video.fosdem.org/2015/devroom-virtualisation/observability.mp4

Provision and manage Docker containers with Foreman--Roll out your applications 
in containers with ease 
https://video.fosdem.org/2015/devroom-virtualisation/provisiondocker__CAM_ONLY.mp4

Rocket and the App Container Spec 
https://video.fosdem.org/2015/devroom-virtualisation/rocketspec.mp4

Transplantation of VirtualBox to the NOVA microhypervisor 
https://video.fosdem.org/2015/devroom-virtualisation/transplantation.mp4

Under the hood of Docker Containers--In flight and at rest 
https://video.fosdem.org/2015/devroom-virtualisation/underhoodcontainers__CAM_ONLY.mp4

Surviving the Zombie Apocalypse--Containers, KVM, Xen, and Security 
https://video.fosdem.org/2015/devroom-virtualisation/zombieapocalypse.mp4


Deep Dives and Technical Discussions


Compiling oVirt-guest-agent on FreeBSD 
http://lists.ovirt.org/pipermail/users/2015-February/031052.html

Running oVirt's Guest Agent on Atomic as a Privileged Container 
http://www.projectatomic.io/blog/2015/01/running-ovirt-guest-agent-as-privileged-container/

Installing oVirt 3.5 on CentOS 7 (Hosted Engine) 
https://xrsa.net/2015/02/04/installing-ovirt-3-5-on-centos-7-hosted-engine/

Containerizing the oVirt Optimizer 
https://github.com/rgolangh/Fedora-Dockerfiles/tree/ovirt-optimizer/ovirt-optimizer

Monitor oVirt or libvirt with SNMP and Zabbix http://jensd.be/?p=491

Move a guest from VMWare ESX to oVirt or RHEV http://jensd.be/?p=489

Trying out oVirt's Probabilistic Optimizer 
http://community.redhat.com/blog/2015/02/trying-out-ovirts-probabalistic-optimizer/

oVirt Platform Management with Ansible 
https://docs.ansible.com/ovirt_module.html

Adding a Cockpit to Node 
http://dummdida.tumblr.com/post/111458709540/adding-a-cockpit-to-node

Node goes Anaconda 
http://dummdida.tumblr.com/post/111851478485/node-goes-anaconda

OVIRT engine unable to import old Export storage: How to fix 
http://arm2armcos.blogspot.co.il/2015/02/ovirt-engile-unable-to-import-old.html

Manage oVirt environment 
http://serverfault.com/questions/669978/manage-ovirt-environment

Gluster ovirt integration_gluster_meetup_pune_2015 
http://www.slideshare.net/N.Ramesh/gluster-ovirt-integrationglustermeetuppune2015

-- 
Brian Proffitt

Community Liaison
oVirt
Open Source and Standards, Red Hat - http://community.redhat.com
Phone: +1 574 383 9BKP
IRC: bkp @ OFTC
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users