Re: [ovirt-users] Communication errors between engine and nodes?
Hello! FYI: updated 2 days ago 3 hypervisor from my setup to latest 3.5-patternfly , rebooted nodes, and engine and the error seems to be gone: no longer got heartbeat exeeded. On Tue, Mar 17, 2015 at 11:58 AM, Piotr Kliczewski piotr.kliczew...@gmail.com wrote: Hi Roel, You can change this setting in two ways. - you can update it in db directly as you stated (not recommended) - use engine-config -s vdsHeartbeatInSeconds=20 but prior to running this command you need to update config file /etc/ovirt-engine/engine-config/engine-config.properties with vdsHeartbeatInSeconds.type=Integer. This config value is not exposed by default. Thanks, Piotr On Mon, Mar 16, 2015 at 11:18 PM, Roel de Rooy rder...@motto.nl wrote: HI Piotr, Thanks for your reply! If I would like to change the heartbeat value, do I have to update the value within the vdc_options table directly, or should this be done by another way (e.g. config file)? Regards, Roel -Oorspronkelijk bericht- Van: Piotr Kliczewski [mailto:piotr.kliczew...@gmail.com] Verzonden: maandag 16 maart 2015 12:16 Aan: Roel de Rooy CC: Michal Skrivanek; users@ovirt.org Onderwerp: Re: [ovirt-users] Communication errors between engine and nodes? Unfortunately log entries that you copied give me almost no information about nature of your issue. There are few things that we can do to understand what is going on with your setup. Heartbeat functionality provides means to detect whether we still have connection with a host. By default heartbeat timeout is set to 10 seconds but it can be modified by setting vdsHeartbeatInSeconds. In general whenever there are no incoming responses nor heartbeat frame is not received engine will invalidate the connection and will attempt to recover. If reconnection was successful you want see any other consequences of loosing single heartbeat. I would explore stability of your network so if the network is busy or you loose network packets from time to time this kind of entries in the log are expected. You can increase heatbeat value and see whether it will work better for your env. If you confirm that your network is stable we could explore the issue further by setting debug level logging for your engine to understand exactly how the messages are processes by a host and when we receive responses. On Mon, Mar 16, 2015 at 11:34 AM, Roel de Rooy rder...@motto.nl wrote: Received the heartbeat exeeded continuously this morning (seems to be quiet again for now). VM's still continue to work correctly and the storage domains (NFS shares) are still connected and reachable on the nodes, at the exact time that this issue is happening. Contacted our network engineer to see if he could see a load increase on our network, or could see any latency, errors, etc. Unfortunately he could not detect anything yet (he is still investigating this). I have attached both the engine and vdsm logs Engine.log: 2015-03-16 10:10:10,506 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.ListVDSCommand] (DefaultQuartzScheduler_Worker-45) [6d40f562] Command ListVDSCommand(HostName = HOST, HostId = 3b87597e-081b-4c89-9b1e-cb04203259f5, vds=Host[HOST,3b87597e-081b-4c89-9b1e-cb04203259f5]) execution failed. Exception: VDSNetworkException: VDSGenericException: VDSNetworkException: Heartbeat exeeded 2015-03-16 10:10:10,507 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.SpmStatusVDSCommand] (DefaultQuartzScheduler_Worker-35) [2c53103c] Command SpmStatusVDSCommand(HostName = HOST, HostId = 3b87597e-081b-4c89-9b1e-cb04203259f5, storagePoolId = 124ae76f-8acb-412e-91cc-dff9f6ec665d) execution failed. Exception: VDSNetworkException: VDSGenericException: VDSNetworkException: Heartbeat exeeded 2015-03-16 10:10:10,506 WARN [org.ovirt.vdsm.jsonrpc.client.internal.ResponseWorker] (ResponseWorker) Exception thrown during message processing 2015-03-16 10:10:10,507 WARN [org.ovirt.engine.core.vdsbroker.VdsManager] (DefaultQuartzScheduler_Worker-45) [6d40f562] Host HOST is not responding. It will stay in Connecting state for a grace period of 88 seconds and after that an attempt to fence the host will be issued. 2015-03-16 10:10:10,510 INFO [org.ovirt.engine.core.bll.storage.SetStoragePoolStatusCommand] (DefaultQuartzScheduler_Worker-35) [7e61eee] Running command: SetStoragePoolStatusCommand internal: true. Entities affected : ID: 124ae76f-8acb-412e-91cc-dff9f6ec665d Type: StoragePool 2015-03-16 10:10:10,512 INFO [org.ovirt.engine.core.vdsbroker.storage.StoragePoolDomainHelper] (DefaultQuartzScheduler_Worker-35) [7e61eee] Storage Pool 124ae76f-8acb-412e-91cc-dff9f6ec665d - Updating Storage Domain bfa86142-6f2e-44fe-8a9c-cf4390f3b8ae status from Active to Unknown, reason : null 2015-03-16 10:10:10,513 INFO [org.ovirt.engine.core.vdsbroker.storage.StoragePoolDomainHelper]
Re: [ovirt-users] VM failed to start | Bad volume specification
On Mar 19, 2015, at 03:18 , Punit Dambiwal hypu...@gmail.com wrote: Hi All, Is there any one have any idea about this problem...it seems it's bug either in Ovirt or Glusterfs...that's why no one has the idea about itplease correct me if i am wrong…. Hi, as I said, storage access times out; so it seems to me as a gluster setup problem, the storage domain you have your VMs on is not working… Thanks, michal Thanks, Punit On Wed, Mar 18, 2015 at 5:05 PM, Punit Dambiwal hypu...@gmail.com wrote: Hi Michal, Would you mind to let me know the possible messedup things...i will check and try to resolve itstill i am communicating gluster community to resolve this issue... But in the ovirtgluster setup is quite straightso how come it will be messedup with reboot ?? if it can be messedup with reboot then it seems not good and stable technology for the production storage Thanks, Punit On Wed, Mar 18, 2015 at 3:51 PM, Michal Skrivanek michal.skriva...@redhat.com wrote: On Mar 18, 2015, at 03:33 , Punit Dambiwal hypu...@gmail.com wrote: Hi, Is there any one from community can help me to solve this issue...?? Thanks, Punit On Tue, Mar 17, 2015 at 12:52 PM, Punit Dambiwal hypu...@gmail.com wrote: Hi, I am facing one strange issue with ovirt/glusterfsstill didn't find this issue is related with glusterfs or Ovirt Ovirt :- 3.5.1 Glusterfs :- 3.6.1 Host :- 4 Hosts (Compute+ Storage)...each server has 24 bricks Guest VM :- more then 100 Issue :- When i deploy this cluster first time..it work well for me(all the guest VM created and running successfully)but suddenly one day my one of the host node rebooted and none of the VM can boot up now...and failed with the following error Bad Volume Specification VMId :- d877313c18d9783ca09b62acf5588048 VDSM Logs :- http://ur1.ca/jxabi you've got timeouts while accessing storage…so I guess something got messed up on reboot, it may also be just a gluster misconfiguration… Engine Logs :- http://ur1.ca/jxabv [root@cpu01 ~]# vdsClient -s 0 getVolumeInfo e732a82f-bae9-4368-8b98-dedc1c3814de 0002-0002-0002-0002-0145 6d123509-6867-45cf-83a2-6d679b77d3c5 9030bb43-6bc9-462f-a1b9-f6d5a02fb180 status = OK domain = e732a82f-bae9-4368-8b98-dedc1c3814de capacity = 21474836480 voltype = LEAF description = parent = ---- format = RAW image = 6d123509-6867-45cf-83a2-6d679b77d3c5 uuid = 9030bb43-6bc9-462f-a1b9-f6d5a02fb180 disktype = 2 legality = LEGAL mtime = 0 apparentsize = 21474836480 truesize = 4562972672 type = SPARSE children = [] pool = ctime = 1422676305 - I opened same thread earlier but didn't get any perfect answers to solve this issue..so i reopen it... https://www.mail-archive.com/users@ovirt.org/msg25011.html Thanks, Punit ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Windows Server 2012 R2 guest + Memory Ballooning = High CPU
On 03/12/2015 05:34 PM, Doron Fediuck wrote: On 11/03/15 12:38, Winfried de Heiden wrote: Hi All, Running Windows 2012 R2 (64 bit) on oVirt 3.5.1-1.el6 caused a rather high cpu load, even ~20% when doing nothing. Turning off theMemory Balloon Device made the problem disappear. The guest tools are installed using the ovirt-guest-tools-3.5_5.iso. Anyone got the same issue? What is causing this? Seems like a bug to me. Yes it does indeed seem to be a bug, and from how this sounds like, I would suspect the windows ballooning driver. Kind regards, Winny Hi Winny, Can you please open a bug with the relevant versions you have? Thanks, Doron ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Regards, Vinzenz Feenstra | Senior Software Engineer RedHat Engineering Virtualization R D Phone: +420 532 294 625 IRC: vfeenstr or evilissimo Better technology. Faster innovation. Powered by community collaboration. See how it works at redhat.com ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Power Management config on Ovirt
Hi Eli/Martin, On each host power management I have given the other hosts iRMC IPMI address. Is that correct or I have to give same hosts management IP? Ovirt engine is installed on another machine and it can reach the iRMC management IP of each host. My setup is as below. 2 hosts with 6 NICs configured as bond0 on each host. Another IP given for iRMC management port on each host and this IP is configured on the other host power management. VM stored on glusterfs storage and engine installed on another server. I can able to create VM and manually migrate but if one host is down, VM is not migrating to other host. What do I add more to achieve this. Regards Renchu Mathew -Original Message- From: Eli Mesika [mailto:emes...@redhat.com] Sent: Wednesday, March 18, 2015 5:51 PM To: Renchu Mathew Cc: Martin Perina; users@ovirt.org Subject: Re: [ovirt-users] Power Management config on Ovirt - Original Message - From: Renchu Mathew ren...@cracknell.com To: Eli Mesika emes...@redhat.com Cc: Martin Perina mper...@redhat.com, users@ovirt.org Sent: Wednesday, March 18, 2015 3:15:40 PM Subject: RE: [ovirt-users] Power Management config on Ovirt Hi Eli, Those 2 hosts are connected with Fujitsu iRMC management port and power management is configured with ipmi. So it can use this connection to fence the other node, is it correct? No, keep in mind that the one that communicates with the proxy host is the oVirt engine, so , if it is not accessable, oVirt engine can not use it Regards Renchu Mathew -Original Message- From: Eli Mesika [mailto:emes...@redhat.com] Sent: Wednesday, March 18, 2015 4:31 PM To: Renchu Mathew Cc: Martin Perina; users@ovirt.org Subject: Re: [ovirt-users] Power Management config on Ovirt - Original Message - From: Renchu Mathew ren...@cracknell.com To: Martin Perina mper...@redhat.com Cc: users@ovirt.org Sent: Wednesday, March 18, 2015 2:24:46 PM Subject: Re: [ovirt-users] Power Management config on Ovirt Hi Martin, My setup meets all those requirements and I can able to migrate the VM from one host to another manually. Once network cable is pulled off from one of the server other server is also shuts down. Hi Sorry for jumping in late , Yesterday was an election day in ISRAEL... If the other server shuts down when you plug-off the first one and you have only 2 hosts then no fencing will take place since there is no available proxy host to perform the operation Regards Renchu Mathew | Sr. IT Administrator CRACKNELL DUBAI | P.O. Box 66231 | United Arab Emirates | T +971 4 3445417 | F +971 4 3493675 | M +971 50 7386484 ABU DHABI | DUBAI | LONDON | MUSCAT | DOHA | JEDDAH EMAIL ren...@cracknell.com | WEB www.cracknell.com This email, its content and any files transmitted with it are intended solely for the addressee(s) and may be legally privileged and/or confidential. If you are not the intended recipient please let us know by email reply and delete it from the system. Please note that any views or opinions presented in this email do not necessarily represent those of the company. Email transmissions cannot be guaranteed to be secure or error-free as information could be intercepted, corrupted, lost, destroyed, arrive late or incomplete, or contain viruses. The company therefore does not accept liability for any errors or omissions in the contents of this message which arise as a result of email transmission. -Original Message- From: Martin Perina [mailto:mper...@redhat.com] Sent: Tuesday, March 17, 2015 8:31 PM To: Renchu Mathew Cc: users@ovirt.org Subject: Re: [ovirt-users] Power Management config on Ovirt Hi, prior to the test I would check this: - Data Center status is Up - All hosts status is Up - All storage domains status is Up - VM is running If this is valid, you can start your fence testing. But bear in mind what I sent you in previous email: at least one host in DC should be fully functional to be able to fence non responsive host. Martin Perina - Original Message - From: Renchu Mathew ren...@cracknell.com To: Martin Perina mper...@redhat.com Cc: users@ovirt.org Sent: Tuesday, March 17, 2015 5:03:53 PM Subject: RE: [ovirt-users] Power Management config on Ovirt Hi Martin Yes, my test VM still running on this storage. Is it possible to do remote session and check this? Regards Renchu Mathew -Original Message- From: Martin Perina [mailto:mper...@redhat.com] Sent: Tuesday, March 17, 2015 7:30 PM To: Renchu Mathew Cc: users@ovirt.org Subject: Re: [ovirt-users] Power Management config on Ovirt Hi, this is what happened (at least what I was able to read from log): 18:18:02 - host node02 changed status to Connecting -
Re: [ovirt-users] VM failed to start | Bad volume specification
Hi Joop, No split brainbut i have find some disconnecting error in the brick logs :- --- [2015-03-19 09:47:38.201769] I [login.c:82:gf_auth] 0-auth/login: allowed user names: 6ba6d773-cf70-4ba9-98d3-d2e31bcbb14c [2015-03-19 09:47:38.201819] I [server-handshake.c:585:server_setvolume] 0-ds01-server: accepted client from cpu01-13039-2015/03/19-09:47:37:502523-ds01-client-0-0-0 (version: 3.6.1) [2015-03-19 09:47:38.243512] I [server.c:518:server_rpc_notify] 0-ds01-server: disconnecting connection from cpu01-13039-2015/03/19-09:47:37:502523-ds01-client-0-0-0 [2015-03-19 09:47:38.243579] I [client_t.c:417:gf_client_unref] 0-ds01-server: Shutting down connection cpu01-13039-2015/03/19-09:47:37:502523-ds01-client-0-0-0 - On Thu, Mar 19, 2015 at 5:42 PM, Joop jvdw...@xs4all.nl wrote: Please keep it on the list. On 19-3-2015 10:10, Punit Dambiwal wrote: Hi Joop, How i can check the split-brain or not ???i already reblance the storage nodes and also run the heal command for the consistency gluster volume heal YourVolume info split-brain Regards, Joop ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] VM failed to start | Bad volume specification
Hi Michal, The Storage domain is up and running and mounted on all the host nodes...as i updated before that it was working perfectly before but just after reboot can not make the VM poweron... [image: Inline image 1] [image: Inline image 2] [root@cpu01 log]# gluster volume info Volume Name: ds01 Type: Distributed-Replicate Volume ID: 369d3fdc-c8eb-46b7-a33e-0a49f2451ff6 Status: Started Number of Bricks: 48 x 2 = 96 Transport-type: tcp Bricks: Brick1: cpu01:/bricks/1/vol1 Brick2: cpu02:/bricks/1/vol1 Brick3: cpu03:/bricks/1/vol1 Brick4: cpu04:/bricks/1/vol1 Brick5: cpu01:/bricks/2/vol1 Brick6: cpu02:/bricks/2/vol1 Brick7: cpu03:/bricks/2/vol1 Brick8: cpu04:/bricks/2/vol1 Brick9: cpu01:/bricks/3/vol1 Brick10: cpu02:/bricks/3/vol1 Brick11: cpu03:/bricks/3/vol1 Brick12: cpu04:/bricks/3/vol1 Brick13: cpu01:/bricks/4/vol1 Brick14: cpu02:/bricks/4/vol1 Brick15: cpu03:/bricks/4/vol1 Brick16: cpu04:/bricks/4/vol1 Brick17: cpu01:/bricks/5/vol1 Brick18: cpu02:/bricks/5/vol1 Brick19: cpu03:/bricks/5/vol1 Brick20: cpu04:/bricks/5/vol1 Brick21: cpu01:/bricks/6/vol1 Brick22: cpu02:/bricks/6/vol1 Brick23: cpu03:/bricks/6/vol1 Brick24: cpu04:/bricks/6/vol1 Brick25: cpu01:/bricks/7/vol1 Brick26: cpu02:/bricks/7/vol1 Brick27: cpu03:/bricks/7/vol1 Brick28: cpu04:/bricks/7/vol1 Brick29: cpu01:/bricks/8/vol1 Brick30: cpu02:/bricks/8/vol1 Brick31: cpu03:/bricks/8/vol1 Brick32: cpu04:/bricks/8/vol1 Brick33: cpu01:/bricks/9/vol1 Brick34: cpu02:/bricks/9/vol1 Brick35: cpu03:/bricks/9/vol1 Brick36: cpu04:/bricks/9/vol1 Brick37: cpu01:/bricks/10/vol1 Brick38: cpu02:/bricks/10/vol1 Brick39: cpu03:/bricks/10/vol1 Brick40: cpu04:/bricks/10/vol1 Brick41: cpu01:/bricks/11/vol1 Brick42: cpu02:/bricks/11/vol1 Brick43: cpu03:/bricks/11/vol1 Brick44: cpu04:/bricks/11/vol1 Brick45: cpu01:/bricks/12/vol1 Brick46: cpu02:/bricks/12/vol1 Brick47: cpu03:/bricks/12/vol1 Brick48: cpu04:/bricks/12/vol1 Brick49: cpu01:/bricks/13/vol1 Brick50: cpu02:/bricks/13/vol1 Brick51: cpu03:/bricks/13/vol1 Brick52: cpu04:/bricks/13/vol1 Brick53: cpu01:/bricks/14/vol1 Brick54: cpu02:/bricks/14/vol1 Brick55: cpu03:/bricks/14/vol1 Brick56: cpu04:/bricks/14/vol1 Brick57: cpu01:/bricks/15/vol1 Brick58: cpu02:/bricks/15/vol1 Brick59: cpu03:/bricks/15/vol1 Brick60: cpu04:/bricks/15/vol1 Brick61: cpu01:/bricks/16/vol1 Brick62: cpu02:/bricks/16/vol1 Brick63: cpu03:/bricks/16/vol1 Brick64: cpu04:/bricks/16/vol1 Brick65: cpu01:/bricks/17/vol1 Brick66: cpu02:/bricks/17/vol1 Brick67: cpu03:/bricks/17/vol1 Brick68: cpu04:/bricks/17/vol1 Brick69: cpu01:/bricks/18/vol1 Brick70: cpu02:/bricks/18/vol1 Brick71: cpu03:/bricks/18/vol1 Brick72: cpu04:/bricks/18/vol1 Brick73: cpu01:/bricks/19/vol1 Brick74: cpu02:/bricks/19/vol1 Brick75: cpu03:/bricks/19/vol1 Brick76: cpu04:/bricks/19/vol1 Brick77: cpu01:/bricks/20/vol1 Brick78: cpu02:/bricks/20/vol1 Brick79: cpu03:/bricks/20/vol1 Brick80: cpu04:/bricks/20/vol1 Brick81: cpu01:/bricks/21/vol1 Brick82: cpu02:/bricks/21/vol1 Brick83: cpu03:/bricks/21/vol1 Brick84: cpu04:/bricks/21/vol1 Brick85: cpu01:/bricks/22/vol1 Brick86: cpu02:/bricks/22/vol1 Brick87: cpu03:/bricks/22/vol1 Brick88: cpu04:/bricks/22/vol1 Brick89: cpu01:/bricks/23/vol1 Brick90: cpu02:/bricks/23/vol1 Brick91: cpu03:/bricks/23/vol1 Brick92: cpu04:/bricks/23/vol1 Brick93: cpu01:/bricks/24/vol1 Brick94: cpu02:/bricks/24/vol1 Brick95: cpu03:/bricks/24/vol1 Brick96: cpu04:/bricks/24/vol1 Options Reconfigured: diagnostics.count-fop-hits: on diagnostics.latency-measurement: on nfs.disable: on user.cifs: enable auth.allow: 10.10.0.* performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server storage.owner-uid: 36 storage.owner-gid: 36 server.allow-insecure: on network.ping-timeout: 100 [root@cpu01 log]# - [root@cpu01 log]# gluster volume status Status of volume: ds01 Gluster process PortOnline Pid -- Brick cpu01:/bricks/1/vol1 49152 Y 33474 Brick cpu02:/bricks/1/vol1 49152 Y 40717 Brick cpu03:/bricks/1/vol1 49152 Y 18080 Brick cpu04:/bricks/1/vol1 49152 Y 40447 Brick cpu01:/bricks/2/vol1 49153 Y 33481 Brick cpu02:/bricks/2/vol1 49153 Y 40724 Brick cpu03:/bricks/2/vol1 49153 Y 18086 Brick cpu04:/bricks/2/vol1 49153 Y 40453 Brick cpu01:/bricks/3/vol1 49154 Y 33489 Brick cpu02:/bricks/3/vol1 49154 Y 40731 Brick cpu03:/bricks/3/vol1 49154 Y 18097 Brick cpu04:/bricks/3/vol1 49154 Y 40460
Re: [ovirt-users] virtio-blk-data-plane
On Oct 27, 2014, at 09:09 , Wolfgang Bucher wolfgang.buc...@netland-mn.de wrote: Hello, is it possible to use virtio-blk-data-plane in ovirt with el7 host. Hi, no, not yet, planned for 3.6 though. Thanks, michal Thanks Wolfgang ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] VM failed to start | Bad volume specification
Hi Joop, How i can check the split-brain or not ???i already reblance the storage nodes and also run the heal command for the consistency On Thu, Mar 19, 2015 at 4:36 PM, Joop jvdw...@xs4all.nl wrote: On 19-3-2015 9:25, Punit Dambiwal wrote: Hi Michal, The Storage domain is up and running and mounted on all the host nodes...as i updated before that it was working perfectly before but just after reboot can not make the VM poweron... Did you check if you're suffering from a split-brain situation? My test setup sometimes does when not being careful when updating glusterfs/ovirt. Regards, Joop ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] VM failed to start | Bad volume specification
Please keep it on the list. On 19-3-2015 10:10, Punit Dambiwal wrote: Hi Joop, How i can check the split-brain or not ???i already reblance the storage nodes and also run the heal command for the consistency gluster volume heal YourVolume info split-brain Regards, Joop ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] VM failed to start | Bad volume specification
On 19-3-2015 9:25, Punit Dambiwal wrote: Hi Michal, The Storage domain is up and running and mounted on all the host nodes...as i updated before that it was working perfectly before but just after reboot can not make the VM poweron... Did you check if you're suffering from a split-brain situation? My test setup sometimes does when not being careful when updating glusterfs/ovirt. Regards, Joop ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Power Management config on Ovirt
Hi, - Original Message - From: Renchu Mathew ren...@cracknell.com To: Eli Mesika emes...@redhat.com Cc: Martin Perina mper...@redhat.com, users@ovirt.org Sent: Thursday, March 19, 2015 9:30:30 AM Subject: RE: [ovirt-users] Power Management config on Ovirt Hi Eli/Martin, On each host power management I have given the other hosts iRMC IPMI address. Is that correct or I have to give same hosts management IP? Each host should have its own IPMI address assigned in PM settings. For example you have these two hosts: 1. host1 with IP1 and IPMI1 2. host2 with IP2 and IPMI2 So when you define a host in oVirt you should use these values in host configuration: 1. host1 IP Address: IP1 Power Management IP address: IPMI1 2. host2 IP Address: IP2 Power Management IP address: IPMI2 Ovirt engine is installed on another machine and it can reach the iRMC management IP of each host. My setup is as below. This is correct, but it's not necessary for engine machine to access power management IPs. But it's required that each host in the cluster have to be able to access all other host power management interfaces. Engine will select one functional host (fence proxy) and on this host power management action is executed. For example you may have this scenario: 1. host1 is Up, connection to host2 is broken 2. engine will select host1 as fence proxy 3. On host1 power management commands are executed (status, stop, start) 2 hosts with 6 NICs configured as bond0 on each host. Another IP given for iRMC management port on each host and this IP is configured on the other host power management. VM stored on glusterfs storage and engine installed on another server. I can able to create VM and manually migrate but if one host is down, VM is not migrating to other host. What do I add more to achieve this. It looks OK except for power management IP address, see above. Martin Regards Renchu Mathew -Original Message- From: Eli Mesika [mailto:emes...@redhat.com] Sent: Wednesday, March 18, 2015 5:51 PM To: Renchu Mathew Cc: Martin Perina; users@ovirt.org Subject: Re: [ovirt-users] Power Management config on Ovirt - Original Message - From: Renchu Mathew ren...@cracknell.com To: Eli Mesika emes...@redhat.com Cc: Martin Perina mper...@redhat.com, users@ovirt.org Sent: Wednesday, March 18, 2015 3:15:40 PM Subject: RE: [ovirt-users] Power Management config on Ovirt Hi Eli, Those 2 hosts are connected with Fujitsu iRMC management port and power management is configured with ipmi. So it can use this connection to fence the other node, is it correct? No, keep in mind that the one that communicates with the proxy host is the oVirt engine, so , if it is not accessable, oVirt engine can not use it Regards Renchu Mathew -Original Message- From: Eli Mesika [mailto:emes...@redhat.com] Sent: Wednesday, March 18, 2015 4:31 PM To: Renchu Mathew Cc: Martin Perina; users@ovirt.org Subject: Re: [ovirt-users] Power Management config on Ovirt - Original Message - From: Renchu Mathew ren...@cracknell.com To: Martin Perina mper...@redhat.com Cc: users@ovirt.org Sent: Wednesday, March 18, 2015 2:24:46 PM Subject: Re: [ovirt-users] Power Management config on Ovirt Hi Martin, My setup meets all those requirements and I can able to migrate the VM from one host to another manually. Once network cable is pulled off from one of the server other server is also shuts down. Hi Sorry for jumping in late , Yesterday was an election day in ISRAEL... If the other server shuts down when you plug-off the first one and you have only 2 hosts then no fencing will take place since there is no available proxy host to perform the operation Regards Renchu Mathew | Sr. IT Administrator CRACKNELL DUBAI | P.O. Box 66231 | United Arab Emirates | T +971 4 3445417 | F +971 4 3493675 | M +971 50 7386484 ABU DHABI | DUBAI | LONDON | MUSCAT | DOHA | JEDDAH EMAIL ren...@cracknell.com | WEB www.cracknell.com This email, its content and any files transmitted with it are intended solely for the addressee(s) and may be legally privileged and/or confidential. If you are not the intended recipient please let us know by email reply and delete it from the system. Please note that any views or opinions presented in this email do not necessarily represent those of the company. Email transmissions cannot be guaranteed to be secure or error-free as information could be intercepted, corrupted, lost, destroyed, arrive late or incomplete, or contain viruses. The company therefore does not accept liability for any errors or omissions in the contents of this message which arise as a result of email transmission. -Original Message- From: Martin Perina
Re: [ovirt-users] 答复: bonding 802.3ad mode
On Wed, Mar 18, 2015 at 10:57:18PM -0400, Dan Yasny wrote: Mode 0 is not supported under a bridge, just like mode 6 On Wed, Mar 18, 2015 at 10:47 PM, Xie, Chao xiec.f...@cn.fujitsu.com wrote: Yeah, Alex is right. And if you want to double the network’s speed in single flow, the mode 0 is only choice. But mode 0 seems not be supported in oVirt? *发件人:* users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] *代表 *Alex Crow *发送时间:* 2015年3月19日 0:25 *收件人:* users@ovirt.org *主题:* Re: [ovirt-users] bonding 802.3ad mode The balancing on 802.3ad only occurs for different network flows based on a hash of source and destination MAC (or can be made to add IP addresses into the calculation). A single flow will only use a single NIC in ad mode. Alex On 18/03/15 16:17, Nathanaël Blanchet wrote: Hi all, I'm used to create a mode 4 bond0 interface with two 1 Gb/s interfaces on all my hosts, and ethtool bond0 gives me a functionnal 2000Mb/s. However, when importing a vm from the export domain (NFS with a speed of 4GB/s), I always have this alert: Host siple has network interface which exceeded the defined threshold [95%] (em3: transmit rate[0%], receive rate [100%]) It seems that the second nic never works while the first one is overloaded. Is it an expected behaviour? I believed that the flow was balanced between the two interfaces in 802.3ad mode. To follow up on former ressponses: what do you have on top of your bond? If you have a VM network, multiple guests are expected to have a different hash value for each, and to spread the load on mode 4. If you use the bonds for a host network (e.g. dispaly, migration, storage) you can try mode 0. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] test, please ignore
A post didn't reach the list, this is a test, please ignore. Mike ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] bonding 802.3ad mode
Bonjour Nathanael, You haven't mentioned which version of oVirt you were using - I suspect it's pre-3.5 and therefore this isn't fixed yet: https://bugzilla.redhat.com/show_bug.cgi?id=1114085 From 3.5 onwards the warning should not appear for a bond (in aggregating mode) if only one of its slaves is overloaded. A bientot, Lior. On 18/03/15 18:17, Nathanaël Blanchet wrote: Hi all, I'm used to create a mode 4 bond0 interface with two 1 Gb/s interfaces on all my hosts, and ethtool bond0 gives me a functionnal 2000Mb/s. However, when importing a vm from the export domain (NFS with a speed of 4GB/s), I always have this alert: Host siple has network interface which exceeded the defined threshold [95%] (em3: transmit rate[0%], receive rate [100%]) It seems that the second nic never works while the first one is overloaded. Is it an expected behaviour? I believed that the flow was balanced between the two interfaces in 802.3ad mode. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Hosted Engine config in our hardware environment
Hello oVirt guru out there: I want to seek some advice on upgrade path for our oVirt management vm configuration. We have been using oVirt for over 3 years. When we first setup oVirt environment, Hosted Engine componment did not exist. Our question is should we migrate our current configuration to use Hosted Engine? First let me give an overview of our configuration. We have blade servers in 2 separate racks. D1 X1. Each side has 10 blades. Storage is iSCSI SAN. Inside our oVirt 3.5.0.1-1.el6 installation, it is configured with 2 data centers. D1 X1. Each datacenter has the 10 blades for that side. The management function of oVirt (oVirt web console) is running off 2 VMs, ovirt-mgmt-1 on D1, and ovirt-mgmt-2 on X1. We have keepalived to maintain a flowing IP for the oVirt management console. The keepalived script makes sure only one copy of ovirt-engine is running at any time. It can be on D1 or X1. The mgmt VMs have Postgresql setup in replication mode. In case one of the mgmt vm failed, the other mgmt vm on the other rack can pick up the mgmt role. Both mgmt VMs can see all blades and SAN resources on D1 X1. This configuration has been working well for us. The drawback is if both ovirt mgmt vm crashed, we will not be able to start them or make any change to the ovirt environment. It is because the mgmt VMs are running within the oVirt domain. We tried to upgrade our configuration to Hosted Engine configuration. From what I understand, the Hosted Engine will run in a separate storage domain. In both times we tried to upgrade to Hosted Engine, they both failed during export and import of current configuration. I think my questions are: - will the Hosted Engine model works in our hardware configuration. With hardware in 2 racks, D1 X1. Can a single Hosted Engine manage hardware on both sides? - How can we achieve redundancy when running Hosted Engine? We need to have management capibility in case one side completely failed. Thanks in advance. Eric ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Cluster comms error, can't migrate
Hi Everyone I have a two node hosted engine cluster that's been running for a month or two. NFS is used for the VM's shared off the nodes on a second network interface with different hostnames, I hope easier to migrate later on. NFS 172.16.67.0/24 ov1-nfs.domain.dom on .1 and ov2-nfs.domain.dom on .2. The NFS shares are working. Management net is 10.10.10.224/28 Last night the cluster had communication errors, but I could not find any issues, all nodes can ping ssh with each other and engine. Today, it got worse, the engine migrated all but 3 VM's to OV2, the node hosting the engine. The VMs still on OV1 are there because the migration for those failed. I can't manually migrate anything back to ov1. I eventually shut down the engine and started on OV1, but still no joy. The VMs are alive, both on OV1 OV2. OV2 is currently in local maintenance to stop the engine moving and stop the email alerts. I have been through the logs, I see there be a cert issue in libvirtd.log on the receiving host? The web interface cert has been changed for a CA issued one several weeks ago, this is working and I monitor the manager using nagios via the API. This is still working. I did not change any of the other internal oVirt PKI stuff. The portal is working to view the VMs and config, but I haven't tried changing anything during this issue apart from trying the migrate. The VMs are providing public services, so I have not tried changing things. Linux ov1.domain.dom 2.6.32-504.3.3.el6.x86_64 #1 SMP Wed Dec 17 01:55:02 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux oVirt Engine Version: 3.5.1-1.el6 OS Version: RHEL - 6 - 6.el6.centos.12.2 Kernel Version: 2.6.32 - 504.3.3.el6.x86_64 KVM Version: 0.12.1.2 - 2.448.el6_6 LIBVIRT Version: libvirt-0.10.2-46.el6_6.2 VDSM Version: vdsm-4.16.10-8.gitc937927.el6 SPICE Version: 0.12.4 - 11.el6 Any help appreciated. Mike [root@ov1 ~]# libvirtd.log 2015-03-18 15:42:17.387+: 3017: error : virNetTLSContextValidCertificate:1008 : Unable to verify TLS peer: The peer did not send any certificate. 2015-03-18 15:42:17.387+: 3017: warning : virNetTLSContextCheckCertificate:1142 : Certificate check failed Unable to verify TLS peer: The peer did not send any certificate. 2015-03-18 15:42:17.387+: 3017: error : virNetTLSContextCheckCertificate:1145 : authentication failed: Failed to verify peer's certificate [root@ov2 ~]# vdsm.log Thread-49490::DEBUG::2015-03-18 15:42:17,294::migration::298::vm.Vm::(_startUnderlyingMigration) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::starting migration to qemu +tls://ov1.domain.dom/system with miguri tcp://10.10.10.227 Thread-49525::DEBUG::2015-03-18 15:42:17,296::migration::361::vm.Vm::(run) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::migration downtime thread started Thread-49526::DEBUG::2015-03-18 15:42:17,297::migration::410::vm.Vm::(monitor_migration) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::starting migration monitor thread Thread-49490::DEBUG::2015-03-18 15:42:17,388::libvirtconnection::143::root::(wrapper) Unknown libvirterror: ecode: 9 edom: 10 level: 2 message: operation failed: Failed to connect to remote libvirt URI qemu+tls://ov1.domain.dom/system Thread-49490::DEBUG::2015-03-18 15:42:17,390::migration::376::vm.Vm::(cancel) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::canceling migration downtime thread Thread-49525::DEBUG::2015-03-18 15:42:17,391::migration::373::vm.Vm::(run) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::migration downtime thread exiting Thread-49490::DEBUG::2015-03-18 15:42:17,391::migration::470::vm.Vm::(stop) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::stopping migration monitor thread Thread-49490::ERROR::2015-03-18 15:42:17,393::migration::161::vm.Vm::(_recover) vmId=`b44b2182-f943-4987-8421-8a98fd2a04d4`::operation failed: Failed to connect to remote libvirt URI qemu+tls://ov1.domain.dom/system [root@ov1 ~]# cat /var/log/vdsm/vdsm.log|grep MY_VM Thread-7589263::DEBUG::2015-03-18 15:22:01,936::BindingXMLRPC::1133::vds::(wrapper) client [10.10.10.228]::call vmMigrationCreate with ({'status': 'Up', 'acpiEnable': 'true', 'emulatedMachine': 'rhel6.5.0', 'afterMigrationStatus': '', 'tabletEnable': 'true', 'vmId': 'b44b2182-f943-4987-8421-8a98fd2a04d4', 'memGuaranteedSize': 2048, 'transparentHugePages': 'true', 'displayPort': '5929', 'displaySecurePort': '-1', 'spiceSslCipherSuite': 'DEFAULT', 'cpuType': 'SandyBridge', 'smp': '2', 'migrationDest': 'libvirt', 'custom': {}, 'vmType': 'kvm', '_srcDomXML': domain type='kvm' id='58'\n nameMY_VM/name\n uuidb44b2182-f943-4987-8421-8a98fd2a04d4/uuid\n memory unit='KiB'2097152/memory\n currentMemory unit='KiB'2097152/currentMemory\n vcpu placement='static' current='2'16/vcpu\n cputune\nshares1020/shares\n /cputune\n sysinfo type='smbios'\nsystem\n entry name='manufacturer'oVirt/entry\n entry name='product'oVirt Node /entry\n entry name='version'6-6.el6.centos.12.2/entry\n entry
[ovirt-users] oVirt Newsletter Special Edition: Mid-March 2015
It's a nice problem to have: it's been so busy in the oVirt community, we have to push out a mid-month newsletter just to keep up! In the Community Solving (NP-Hard) Scheduling Problems with oVirt and OptaPlanner http://www.socallinuxexpo.org/scale/13x/presentations/solving-np-hard-scheduling-problems-ovirt-and-optaplanner FOSDEM Virtualization Room Videos: Docker Integration in oVirt and IaaS https://video.fosdem.org/2015/devroom-virtualisation/dockerovirt__CAM_ONLY.mp4 Live migration for containers is around the corner https://video.fosdem.org/2015/devroom-virtualisation/livemigration__CAM_ONLY.mp4 LXD: The Container Hypervisor https://video.fosdem.org/2015/devroom-virtualisation/lxd.mp4 Observability in KVM--How to troubleshoot virtual machines https://video.fosdem.org/2015/devroom-virtualisation/observability.mp4 Provision and manage Docker containers with Foreman--Roll out your applications in containers with ease https://video.fosdem.org/2015/devroom-virtualisation/provisiondocker__CAM_ONLY.mp4 Rocket and the App Container Spec https://video.fosdem.org/2015/devroom-virtualisation/rocketspec.mp4 Transplantation of VirtualBox to the NOVA microhypervisor https://video.fosdem.org/2015/devroom-virtualisation/transplantation.mp4 Under the hood of Docker Containers--In flight and at rest https://video.fosdem.org/2015/devroom-virtualisation/underhoodcontainers__CAM_ONLY.mp4 Surviving the Zombie Apocalypse--Containers, KVM, Xen, and Security https://video.fosdem.org/2015/devroom-virtualisation/zombieapocalypse.mp4 Deep Dives and Technical Discussions Compiling oVirt-guest-agent on FreeBSD http://lists.ovirt.org/pipermail/users/2015-February/031052.html Running oVirt's Guest Agent on Atomic as a Privileged Container http://www.projectatomic.io/blog/2015/01/running-ovirt-guest-agent-as-privileged-container/ Installing oVirt 3.5 on CentOS 7 (Hosted Engine) https://xrsa.net/2015/02/04/installing-ovirt-3-5-on-centos-7-hosted-engine/ Containerizing the oVirt Optimizer https://github.com/rgolangh/Fedora-Dockerfiles/tree/ovirt-optimizer/ovirt-optimizer Monitor oVirt or libvirt with SNMP and Zabbix http://jensd.be/?p=491 Move a guest from VMWare ESX to oVirt or RHEV http://jensd.be/?p=489 Trying out oVirt's Probabilistic Optimizer http://community.redhat.com/blog/2015/02/trying-out-ovirts-probabalistic-optimizer/ oVirt Platform Management with Ansible https://docs.ansible.com/ovirt_module.html Adding a Cockpit to Node http://dummdida.tumblr.com/post/111458709540/adding-a-cockpit-to-node Node goes Anaconda http://dummdida.tumblr.com/post/111851478485/node-goes-anaconda OVIRT engine unable to import old Export storage: How to fix http://arm2armcos.blogspot.co.il/2015/02/ovirt-engile-unable-to-import-old.html Manage oVirt environment http://serverfault.com/questions/669978/manage-ovirt-environment Gluster ovirt integration_gluster_meetup_pune_2015 http://www.slideshare.net/N.Ramesh/gluster-ovirt-integrationglustermeetuppune2015 -- Brian Proffitt Community Liaison oVirt Open Source and Standards, Red Hat - http://community.redhat.com Phone: +1 574 383 9BKP IRC: bkp @ OFTC ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users