Re: [Users] Disk Migration
On Feb 27, 2014, at 08:15 , Nir Soffer nsof...@redhat.com wrote: - Original Message - From: Dafna Ron d...@redhat.com To: Maurice James midnightst...@msn.com Cc: Ofer Blaut obl...@redhat.com, users@ovirt.org Sent: Wednesday, February 26, 2014 7:34:11 PM Subject: Re: [Users] Disk Migration On 02/26/2014 05:24 PM, Maurice James wrote: I have a specific interface set up for migrations. Why do disk migrations not use the interface that I have set for migrations? Is that by design? Shouldnt it use the interfaces that I have set aside for migrations? VM migrations work as they should but not disk migrations I don't think that you can configure interface for disk migration. Disk migration is actually copy of information from the original disk to a new disk created on a new domain + delete of the original disk once that is done. it's not actually a migration and so I am not sure you can actually configure an interface for that. adding ofer - perhpas he has a solution or it's possible and I am not aware of it. I guess that *not* using the migration network for storage operation is the expected behavior, to make migration faster and safer. Michal, Dan, can you elaborate on this? with storage offloading it's probably not going to be significant, however today it likely is. Nir, why would not using migration network make it better? Won't we have the same problem as before without migration network at all, i.e. choking the management channel? Should we maybe consider a dedicated storage network? Thanks, michal Nir ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Migration Failed
Hi Koen, Can you please attach the relevant vdsm logs? - Original Message - From: Koen Vanoppen vanoppen.k...@gmail.com To: users@ovirt.org Sent: Thursday, February 27, 2014 9:38:46 AM Subject: [Users] Migration Failed Dear all, I added a new host to our ovirt. Everything went good, exept in the beginnen there was a problem with the firmware of the FibreCard but that is solved (maybe relevant to the issue coming up ;-) ), host is green en up now. But when I tried to migrate a machine for testing purpose to see if everythin was ok, I get the following error in the engine.log and the migration fails: 2014-02-27 08:33:08,082 INFO [org.ovirt.engine.core.bll.MigrateVmCommand] (pool-6-thread-49) [f1a68d8] Running command: MigrateVmCommand internal: false. Entities affected : ID: 3444fc9d-0395-4cbb-9a11-28a42802560c Type: VM 2014-02-27 08:33:08,362 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (pool-6-thread-49) [f1a68d8] START, MigrateVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstVdsId=6707fa40-753a-4c95-9304-e47198477e4d, dstHost= buran.brusselsairport.aero:54321 , migrationMethod=ONLINE, tunnelMigration=false), log id: 50cd7284 2014-02-27 08:33:08,371 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] VdsBroker::migrate::Entered (vm_guid=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstHost= buran.brusselsairport.aero:54321 , method=online 2014-02-27 08:33:08,405 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] START, MigrateBrokerVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstVdsId=6707fa40-753a-4c95-9304-e47198477e4d, dstHost= buran.brusselsairport.aero:54321 , migrationMethod=ONLINE, tunnelMigration=false), log id: 20806b79 2014-02-27 08:33:08,441 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] FINISH, MigrateBrokerVDSCommand, log id: 20806b79 2014-02-27 08:33:08,451 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (pool-6-thread-49) [f1a68d8] FINISH, MigrateVDSCommand, return: MigratingFrom, log id: 50cd7284 2014-02-27 08:33:08,491 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (pool-6-thread-49) [f1a68d8] Correlation ID: f1a68d8, Job ID: c3642418-3f05-41eb-8b1d-07fe04867742, Call Stack: null, Custom Event ID: -1, Message: Migration started (VM: ADW-DevSplunk, Source: soyuz, Destination: buran, User: admin@internal). 2014-02-27 08:33:20,036 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) VM ADW-DevSplunk 3444fc9d-0395-4cbb-9a11-28a42802560c moved from MigratingFrom -- Up 2014-02-27 08:33:20,042 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) Adding VM 3444fc9d-0395-4cbb-9a11-28a42802560c to re-run list 2014-02-27 08:33:20,051 ERROR [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) Rerun vm 3444fc9d-0395-4cbb-9a11-28a42802560c. Called from vds soyuz 2014-02-27 08:33:20,107 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) START, MigrateStatusVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c), log id: 75ac0a46 2014-02-27 08:33:20,124 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Failed in MigrateStatusVDS method 2014-02-27 08:33:20,130 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Error code noConPeer and error message VDSGenericException: VDSErrorException: Failed to MigrateStatusVDS, error = Could not connect to peer VDS 2014-02-27 08:33:20,136 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Command org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand return value StatusOnlyReturnForXmlRpc [mStatus=StatusForXmlRpc [mCode=10, mMessage=Could not connect to peer VDS]] 2014-02-27 08:33:20,139 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) HostName = soyuz 2014-02-27 08:33:20,145 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Command MigrateStatusVDS execution failed. Exception: VDSErrorException: VDSGenericException: VDSErrorException: Failed to MigrateStatusVDS, error = Could not connect to peer VDS 2014-02-27 08:33:20,148 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) FINISH, MigrateStatusVDSCommand, log id: 75ac0a46
Re: [Users] New repositories layout for nightly and official releases
This directory will host released builds as well. So we don't really care about pub name. - Kiril - Original Message - From: Yedidyah Bar David d...@redhat.com To: Kiril Nesenko knese...@redhat.com Cc: users@ovirt.org Sent: Wednesday, February 26, 2014 1:10:48 PM Subject: Re: [Users] New repositories layout for nightly and official releases - Original Message - From: Kiril Nesenko knese...@redhat.com To: annou...@ovirt.org, users@ovirt.org Sent: Wednesday, February 26, 2014 12:17:06 PM Subject: [Users] New repositories layout for nightly and official releases Hello all, oVirt Release Engineering team would like to announce a new repositories layout. Nighlty builds are moving to [1]. We are going to disable nightly builds under [2]. Short description to make things clear: ovirt-version-snapshot - nighltly build for the version ovirt-version-snapshot-static - static repo for the nightly build. Will be used for static rpms like jasper, jboss. ovirt-version - official released version Complete by itself, no need for extra static repo If you have any issues, please drop email to in...@ovirt.org or contact us on IRC. Release Manager: Sandro Bonazzola (sbonazzo) Release Engineering: Kiril Nesenko (knesenko), David Caro (dcaroest) With regards, oVirt Release Engineering [1] http://resources.ovirt.org/pub/ [2] http://resources.ovirt.org/releases/nightly/ Nice! I understand that /pub/ovirt-snapshot/ is built from master. I suggest to rename it to ovirt-master-snapshot to be consistent with the other names. Also, why pub? It sounds as if it's public and the rest isn't. If you do not want to reuse the name nightly, perhaps snapshots? E.g. /snapshots/ovirt-3.3-snapshot/, /snapshots/ovirt-master-snapshot/ -- Didi ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
Well, you generate a separate ticket to the same vm. resulting in another vnc-session to the server, but the input/output is shared, you can see in real time when the other user types etc. Am 26.02.2014 14:15, schrieb Itamar Heim: On 02/26/2014 11:07 AM, Sven Kieske wrote: Well, as I mentioned in the BZ: Shared access does work for novnc console, maybe this is good enough for your needs? how does it work wrt to the ticketing mechanism? -- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
On 02/27/2014 10:51 AM, Sven Kieske wrote: Well, you generate a separate ticket to the same vm. resulting in another vnc-session to the server, but the input/output is shared, you can see in real time when the other user types etc. iirc, the ticketing mechanism works in a way that a new ticket disconnects the previous session? Am 26.02.2014 14:15, schrieb Itamar Heim: On 02/26/2014 11:07 AM, Sven Kieske wrote: Well, as I mentioned in the BZ: Shared access does work for novnc console, maybe this is good enough for your needs? how does it work wrt to the ticketing mechanism? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
On 02/27/2014 10:52 AM, Itamar Heim wrote: On 02/27/2014 10:51 AM, Sven Kieske wrote: Well, you generate a separate ticket to the same vm. resulting in another vnc-session to the server, but the input/output is shared, you can see in real time when the other user types etc. iirc, the ticketing mechanism works in a way that a new ticket disconnects the previous session? (on connection with the new ticket) Am 26.02.2014 14:15, schrieb Itamar Heim: On 02/26/2014 11:07 AM, Sven Kieske wrote: Well, as I mentioned in the BZ: Shared access does work for novnc console, maybe this is good enough for your needs? how does it work wrt to the ticketing mechanism? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Migration Failed
Hi, Unfortunately it seems the vdsm logs cycled - these vdsm logs do not match the times for the engine log snippet you pasted - they start at around 10:00 AM and the engine points to 8:33... Gadi Ickowicz - Original Message - From: Koen Vanoppen vanoppen.k...@gmail.com To: Meital Bourvine mbour...@redhat.com, users@ovirt.org Sent: Thursday, February 27, 2014 11:04:11 AM Subject: Re: [Users] Migration Failed In attachment... Thanx! 2014-02-27 9:21 GMT+01:00 Meital Bourvine mbour...@redhat.com : Hi Koen, Can you please attach the relevant vdsm logs? From: Koen Vanoppen vanoppen.k...@gmail.com To: users@ovirt.org Sent: Thursday, February 27, 2014 9:38:46 AM Subject: [Users] Migration Failed Dear all, I added a new host to our ovirt. Everything went good, exept in the beginnen there was a problem with the firmware of the FibreCard but that is solved (maybe relevant to the issue coming up ;-) ), host is green en up now. But when I tried to migrate a machine for testing purpose to see if everythin was ok, I get the following error in the engine.log and the migration fails: 2014-02-27 08:33:08,082 INFO [org.ovirt.engine.core.bll.MigrateVmCommand] (pool-6-thread-49) [f1a68d8] Running command: MigrateVmCommand internal: false. Entities affected : ID: 3444fc9d-0395-4cbb-9a11-28a42802560c Type: VM 2014-02-27 08:33:08,362 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (pool-6-thread-49) [f1a68d8] START, MigrateVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstVdsId=6707fa40-753a-4c95-9304-e47198477e4d, dstHost= buran.brusselsairport.aero:54321 , migrationMethod=ONLINE, tunnelMigration=false), log id: 50cd7284 2014-02-27 08:33:08,371 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] VdsBroker::migrate::Entered (vm_guid=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstHost= buran.brusselsairport.aero:54321 , method=online 2014-02-27 08:33:08,405 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] START, MigrateBrokerVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstVdsId=6707fa40-753a-4c95-9304-e47198477e4d, dstHost= buran.brusselsairport.aero:54321 , migrationMethod=ONLINE, tunnelMigration=false), log id: 20806b79 2014-02-27 08:33:08,441 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] FINISH, MigrateBrokerVDSCommand, log id: 20806b79 2014-02-27 08:33:08,451 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (pool-6-thread-49) [f1a68d8] FINISH, MigrateVDSCommand, return: MigratingFrom, log id: 50cd7284 2014-02-27 08:33:08,491 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (pool-6-thread-49) [f1a68d8] Correlation ID: f1a68d8, Job ID: c3642418-3f05-41eb-8b1d-07fe04867742, Call Stack: null, Custom Event ID: -1, Message: Migration started (VM: ADW-DevSplunk, Source: soyuz, Destination: buran, User: admin@internal). 2014-02-27 08:33:20,036 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) VM ADW-DevSplunk 3444fc9d-0395-4cbb-9a11-28a42802560c moved from MigratingFrom -- Up 2014-02-27 08:33:20,042 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) Adding VM 3444fc9d-0395-4cbb-9a11-28a42802560c to re-run list 2014-02-27 08:33:20,051 ERROR [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) Rerun vm 3444fc9d-0395-4cbb-9a11-28a42802560c. Called from vds soyuz 2014-02-27 08:33:20,107 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) START, MigrateStatusVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c), log id: 75ac0a46 2014-02-27 08:33:20,124 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Failed in MigrateStatusVDS method 2014-02-27 08:33:20,130 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Error code noConPeer and error message VDSGenericException: VDSErrorException: Failed to MigrateStatusVDS, error = Could not connect to peer VDS 2014-02-27 08:33:20,136 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Command org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand return value StatusOnlyReturnForXmlRpc [mStatus=StatusForXmlRpc [mCode=10, mMessage=Could not connect to peer VDS]] 2014-02-27 08:33:20,139 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50)
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
I just tested it again via the implemented novnc solution in ovirt 3.3.3-2 el6. Steps to reproduce: 1. Make a vm accessible via novnc 2. click the console button in webadmin 3. a new browser window appears 4. click the console button again 5. a new browser window appears. 6. put the windows side by side, then type in some words in a shell in window 1 and watch the words also appear in window 2. I don't know if this is a bug, as far as I understand websocketproxy and novnc there is no way how the websocketproxy could now about existing sessions as it forks a subprocess per each session, so it must keep a database or similar to track which session is which subprocess. but even this wouldn't work as the websocketproxy just makes the connection to the compute node hosting the vm, the websocketproxy does not now to which vm you connect (afaik). Is this a bug or a feature? Try it yourself. Am 27.02.2014 09:52, schrieb Itamar Heim: iirc, the ticketing mechanism works in a way that a new ticket disconnects the previous session? -- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
Von: users-boun...@ovirt.org [users-boun...@ovirt.org]quot; im Auftrag von quot;Itamar Heim [ih...@redhat.com] Gesendet: Donnerstag, 27. Februar 2014 09:53 An: Sven Kieske; users@ovirt.org Betreff: Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning] On 02/27/2014 10:52 AM, Itamar Heim wrote: On 02/27/2014 10:51 AM, Sven Kieske wrote: Well, you generate a separate ticket to the same vm. resulting in another vnc-session to the server, but the input/output is shared, you can see in real time when the other user types etc. iirc, the ticketing mechanism works in a way that a new ticket disconnects the previous session? (on connection with the new ticket) In the VNC case this does not work. I do not know if its by design or just a bug/feature. Either Ovirt does not instruct libvirt to create a new VNC ticket with option disconnect all connected clients or this information is not correctly interpreted by qemu. If you launch a client that initiates a shared connection to the console then existing connections will stay open. This magically works if you use the novnc client (as Sven explained). The only thing that prevents a concurrent access to the console with virt-viewer is the fact that it enforces a single VNC connection. So state is as follows: SPICE: disconnect happens during ticket creation VNC: disconnect may happen during client connection You can check that behaviour if you associate a .VV file with a text editor. With SPICE the existing console connection is terminated just before the .VV file opens in the editor. With VNC the existing session stays open. Markus Diese E-Mail enthält vertrauliche und/oder rechtlich geschützte Informationen. Wenn Sie nicht der richtige Adressat sind oder diese E-Mail irrtümlich erhalten haben, informieren Sie bitte sofort den Absender und vernichten Sie diese Mail. Das unerlaubte Kopieren sowie die unbefugte Weitergabe dieser Mail ist nicht gestattet. Ãber das Internet versandte E-Mails können unter fremden Namen erstellt oder manipuliert werden. Deshalb ist diese als E-Mail verschickte Nachricht keine rechtsverbindliche Willenserklärung. Collogia Unternehmensberatung AG Ubierring 11 D-50678 Köln Vorstand: Kadir Akin Dr. Michael Höhnerbach Vorsitzender des Aufsichtsrates: Hans Kristian Langva Registergericht: Amtsgericht Köln Registernummer: HRB 52 497 This e-mail may contain confidential and/or privileged information. If you are not the intended recipient (or have received this e-mail in error) please notify the sender immediately and destroy this e-mail. Any unauthorized copying, disclosure or distribution of the material in this e-mail is strictly forbidden. e-mails sent over the internet may have been written under a wrong name or been manipulated. That is why this message sent as an e-mail is not a legally binding declaration of intention. Collogia Unternehmensberatung AG Ubierring 11 D-50678 Köln executive board: Kadir Akin Dr. Michael Höhnerbach President of the supervisory board: Hans Kristian Langva Registry office: district court Cologne Register number: HRB 52 497 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
On Feb 27, 2014, at 10:19 , Markus Stockhausen stockhau...@collogia.de wrote: Von: users-boun...@ovirt.org [users-boun...@ovirt.org]quot; im Auftrag von quot;Itamar Heim [ih...@redhat.com] Gesendet: Donnerstag, 27. Februar 2014 09:53 An: Sven Kieske; users@ovirt.org Betreff: Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning] On 02/27/2014 10:52 AM, Itamar Heim wrote: On 02/27/2014 10:51 AM, Sven Kieske wrote: Well, you generate a separate ticket to the same vm. resulting in another vnc-session to the server, but the input/output is shared, you can see in real time when the other user types etc. iirc, the ticketing mechanism works in a way that a new ticket disconnects the previous session? (on connection with the new ticket) In the VNC case this does not work. I do not know if its by design or just a bug/feature. Either Ovirt does not instruct libvirt to create a new VNC ticket with option disconnect all connected clients or this information is not correctly interpreted by qemu. well, sort of both. We keep the QEMU/libvirt default which is to honor client's request. Which is in case of novnc by default shared, other common clients like vncviewer default to disconnect I wonder if anyone is using the shared behavior or we should disallow it. Or make it optional, which is nice but as always comes with a price of development time. If you launch a client that initiates a shared connection to the console then existing connections will stay open. This magically works if you use the novnc client (as Sven explained). The only thing that prevents a concurrent access to the console with virt-viewer is the fact that it enforces a single VNC connection. yes So state is as follows: SPICE: disconnect happens during ticket creation It's on connection as well, it just can't be controlled (not that it matters:) Thanks, michal VNC: disconnect may happen during client connection You can check that behaviour if you associate a .VV file with a text editor. With SPICE the existing console connection is terminated just before the .VV file opens in the editor. With VNC the existing session stays open. MarkusInterScan_Disclaimer.txt___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Migration Failed
On Feb 27, 2014, at 10:09 , Gadi Ickowicz gicko...@redhat.com wrote: Hi, Unfortunately it seems the vdsm logs cycled - these vdsm logs do not match the times for the engine log snippet you pasted - they start at around 10:00 AM and the engine points to 8:33… seeing error = Could not connect to peer VDS points me to a possible direct network connectivity issue between the those two hosts. Src vdsm needs to be able to talk to dst vdsm Thanks, michal Gadi Ickowicz - Original Message - From: Koen Vanoppen vanoppen.k...@gmail.com To: Meital Bourvine mbour...@redhat.com, users@ovirt.org Sent: Thursday, February 27, 2014 11:04:11 AM Subject: Re: [Users] Migration Failed In attachment... Thanx! 2014-02-27 9:21 GMT+01:00 Meital Bourvine mbour...@redhat.com : Hi Koen, Can you please attach the relevant vdsm logs? From: Koen Vanoppen vanoppen.k...@gmail.com To: users@ovirt.org Sent: Thursday, February 27, 2014 9:38:46 AM Subject: [Users] Migration Failed Dear all, I added a new host to our ovirt. Everything went good, exept in the beginnen there was a problem with the firmware of the FibreCard but that is solved (maybe relevant to the issue coming up ;-) ), host is green en up now. But when I tried to migrate a machine for testing purpose to see if everythin was ok, I get the following error in the engine.log and the migration fails: 2014-02-27 08:33:08,082 INFO [org.ovirt.engine.core.bll.MigrateVmCommand] (pool-6-thread-49) [f1a68d8] Running command: MigrateVmCommand internal: false. Entities affected : ID: 3444fc9d-0395-4cbb-9a11-28a42802560c Type: VM 2014-02-27 08:33:08,362 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (pool-6-thread-49) [f1a68d8] START, MigrateVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstVdsId=6707fa40-753a-4c95-9304-e47198477e4d, dstHost= buran.brusselsairport.aero:54321 , migrationMethod=ONLINE, tunnelMigration=false), log id: 50cd7284 2014-02-27 08:33:08,371 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] VdsBroker::migrate::Entered (vm_guid=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstHost= buran.brusselsairport.aero:54321 , method=online 2014-02-27 08:33:08,405 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] START, MigrateBrokerVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c, srcHost= soyuz.brusselsairport.aero , dstVdsId=6707fa40-753a-4c95-9304-e47198477e4d, dstHost= buran.brusselsairport.aero:54321 , migrationMethod=ONLINE, tunnelMigration=false), log id: 20806b79 2014-02-27 08:33:08,441 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (pool-6-thread-49) [f1a68d8] FINISH, MigrateBrokerVDSCommand, log id: 20806b79 2014-02-27 08:33:08,451 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (pool-6-thread-49) [f1a68d8] FINISH, MigrateVDSCommand, return: MigratingFrom, log id: 50cd7284 2014-02-27 08:33:08,491 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (pool-6-thread-49) [f1a68d8] Correlation ID: f1a68d8, Job ID: c3642418-3f05-41eb-8b1d-07fe04867742, Call Stack: null, Custom Event ID: -1, Message: Migration started (VM: ADW-DevSplunk, Source: soyuz, Destination: buran, User: admin@internal). 2014-02-27 08:33:20,036 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) VM ADW-DevSplunk 3444fc9d-0395-4cbb-9a11-28a42802560c moved from MigratingFrom -- Up 2014-02-27 08:33:20,042 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) Adding VM 3444fc9d-0395-4cbb-9a11-28a42802560c to re-run list 2014-02-27 08:33:20,051 ERROR [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_Worker-82) Rerun vm 3444fc9d-0395-4cbb-9a11-28a42802560c. Called from vds soyuz 2014-02-27 08:33:20,107 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) START, MigrateStatusVDSCommand(HostName = soyuz, HostId = 6dfa2f9c-85c6-4fb3-b65f-c84620115a1a, vmId=3444fc9d-0395-4cbb-9a11-28a42802560c), log id: 75ac0a46 2014-02-27 08:33:20,124 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Failed in MigrateStatusVDS method 2014-02-27 08:33:20,130 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand] (pool-6-thread-50) Error code noConPeer and error message VDSGenericException: VDSErrorException: Failed to MigrateStatusVDS, error = Could not connect to peer VDS 2014-02-27 08:33:20,136 INFO
Re: [Users] concurrent SPICE access [was: Re: oVirt 3.5 planning]
On 02/27/2014 02:30 PM, Michal Skrivanek wrote: well, sort of both. We keep the QEMU/libvirt default which is to honor client's request. Which is in case of novnc by default shared, other common clients like vncviewer default to disconnect I wonder if anyone is using the shared behavior or we should disallow it. Or make it optional, which is nice but as always comes with a price of development time. default should be to dis-allow it, seems there is a use case to let admin set a VM console setting to shared though. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning - bandwidth accounting
+1 This would be very useful! -- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] How do you move an host with local-storage into regular Data Center?
2014-02-26 17:35 GMT+01:00 Dafna Ron d...@redhat.com: you did not remove the storage before moving the host. Well, I tried but then there was always something impossible to do, i.e. the Local Storage Domain was the only SD so, beeng it the Master SD, it wasn't possible to remove it... something like the chicken or the egg dilemma. If you select the force remove DC option it should clean all object under that DC (just make sure you are selecting the one you want to remove ;)) That worked. I was just trying to avoid that forced option, but you confirmed me it was the only one. Thank you, Giorgio. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning - bandwidth accounting
Dan, How about storing the rx_byte per 5 minutes in the engine DB? That way a reset of the counters has a minimal impact and analytics as traffic for VM x in month Y could be made. Another approach could be to have iptables keep the count? On Thu, Feb 27, 2014 at 1:03 PM, Dan Kenigsberg dan...@redhat.com wrote: There are users that would like to tell how much traffic each vnic of each VM has consumed in a period of time. Currently, we report only bitrate as a percetage of an estimated vnic speed. Integrating this value over time is inefficent and error prone. I suggest to have all the stack (Vdsm, Engine, dwh) report the actually-trasmitted (and actually-received) byte count on each vnic, as well as the time when the sample was taken. Currently, Vdsm reports 'eth0': {'rxDropped': '0', 'rxErrors': '0', 'rxRate': '8.0', 'speed': '1000', 'state': 'up', 'txDropped': '0', 'txErrors': '0', 'txRate': '10.0'}, but it should add rxKiBytes, txKiBytes and time to the frill. GUI could still calculate the rate for illustration, based on the raw trasmission and the sample time. Until we break backward compatibility, we'd keep reporting the flaky rxRate/txRate, too. I can think of only two problems with this approach: Linux byte counters would eventually reset when they overflow. This is currently hidden by Vdsm, but with the suggested change, would have to be handled by higher levels of the stack. A similar problem appears on migration: the counters would reset and Engine would need to know how to keep up the accounting properly. I've opened Bug 1066570 - [RFE] Report actual rx_byte instead of a false rxRate to track this request of mine. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Met vriendelijke groeten / With kind regards, Johan Kooijman T +31(0) 6 43 44 45 27 F +31(0) 162 82 00 01 E m...@johankooijman.com ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] gluster
hi, it is my glusterfs log, [2014-02-27 10:44:35.565367] I [client.c:1883:client_rpc_notify] 0-ctdb-client-0: disconnected [2014-02-27 10:44:35.581812] W [socket.c:1494:__socket_proto_state_machine] 0-glusterfs: reading from socket failed. Error (Transport endpoint is not connected), peer (192.168.101.93:24007) Is it harmless or do I need to worry? can you help me,thanks? yfw...@daicy.net___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Problem setting up a ovirt node / engine on a Minimal CentOS 6.5
Hello, I was fortuned to get my hands on a second server with more storage. Now I have a HP Proliant 380 G5 with 2 4-cores and 32 Gb of memory and close to 1,2 Tb of space. Unfortunatly it is equiped with a smart array 400 so I configured it as a RAID 10. (8 300Gb disks) and 6 Gigabyte nic's I would like to set this server up as a ovirt-engine and the NFS server The second server is a DELL PowerEdge R420 with 2 6-cores and 64 Gb RAM and 2 Gigayte nic's My question is how to set up the ovirt engine (partitioning, network setup) and I will use the ovirt-node image you just to install on the DELL server. Kind regards. 2014-02-21 15:18 GMT+01:00 Andy Michielsen andy.michiel...@gmail.com: Hello, Maybe I'm not asking my question at the right guy's but here I go. I installed a clean CentOS 6.5 on a new DELL server and wanted to use it to install oVirt 3.3.x.x I changed the ifcfg-eth0 file to get it connected to the internet and did a yum -y update. Then I added some repositories to get the oVirt packages : - yum -y install http://dl.fedoraproject.org/pub/epel/6/x86_64/epel-release-6-8.noarch.rpm - yum -y install http://resources.ovirt.org/releases/ovirt-release-el.noarch.rpm Installed the ovirt engine first with yum -y install ovirt-engine and than run the setup by engine-setup. So far so good. I can log into the oVirt website and see that there is nothing there. Then I want to install yum -y install vdsm vdsm-cli and wanted to add the ifcfg-ovirtmgmt bridge and that's where thing go wrong. Obviously I did something wrong but I don't know what. Any help would be greatly appriciated. Kind regards. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com To: jo...@israelsson.com; users@ovirt.org; mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users attachment: poweroff.PNG___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] How do you move an host with local-storage into regular Data Center?
Actually to remove the the last domain in a DC the following steps will do it in and clean up everything without resorting to manual steps: 1. Deactivate and then detach all domains in the DC except last one 2. Deactivate last domain 3. Remove the datacenter (this takes a while since actually the host has to contend for SPM again, like reactivating the domain, in order to remove the DC) 4. DC should be removed and domain should remain unattached - now domain can be removed 5. Move host to maintenance and switch it's cluster to proper dc (should now be possible since it's DC has been removed and is now blank) or move the host directly to proper cluster One thing to note when using force remove for DC is that it does *not* clean up the actual storage - only references to it from the DB (and that is why it disappears from the UI). You have to manually go to the host and clean the storage itself to free up that space Gadi Ickowicz - Original Message - From: Giorgio Bersano giorgio.bers...@gmail.com To: d...@redhat.com Cc: users@ovirt.org Users@ovirt.org Sent: Thursday, February 27, 2014 3:04:25 PM Subject: Re: [Users] How do you move an host with local-storage into regular Data Center? 2014-02-26 17:35 GMT+01:00 Dafna Ron d...@redhat.com: you did not remove the storage before moving the host. Well, I tried but then there was always something impossible to do, i.e. the Local Storage Domain was the only SD so, beeng it the Master SD, it wasn't possible to remove it... something like the chicken or the egg dilemma. If you select the force remove DC option it should clean all object under that DC (just make sure you are selecting the one you want to remove ;)) That worked. I was just trying to avoid that forced option, but you confirmed me it was the only one. Thank you, Giorgio. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] How do you move an host with local-storage into regular Data Center?
2014-02-27 14:46 GMT+01:00 Gadi Ickowicz gicko...@redhat.com: Actually to remove the the last domain in a DC the following steps will do it in and clean up everything without resorting to manual steps: 1. Deactivate and then detach all domains in the DC except last one 2. Deactivate last domain 3. Remove the datacenter (this takes a while since actually the host has to contend for SPM again, like reactivating the domain, in order to remove the DC) 4. DC should be removed and domain should remain unattached - now domain can be removed 5. Move host to maintenance and switch it's cluster to proper dc (should now be possible since it's DC has been removed and is now blank) or move the host directly to proper cluster One thing to note when using force remove for DC is that it does *not* clean up the actual storage - only references to it from the DB (and that is why it disappears from the UI). You have to manually go to the host and clean the storage itself to free up that space Gadi Ickowicz Thank you Gadi, I'll try this procedure as soon as I'll have time to experiment. Was it already documented somewhere? In that case please forgive my ignorance. Giorgio ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Disk Migration
On Thu, Feb 27, 2014 at 09:12:26AM +0100, Michal Skrivanek wrote: On Feb 27, 2014, at 08:15 , Nir Soffer nsof...@redhat.com wrote: - Original Message - From: Dafna Ron d...@redhat.com To: Maurice James midnightst...@msn.com Cc: Ofer Blaut obl...@redhat.com, users@ovirt.org Sent: Wednesday, February 26, 2014 7:34:11 PM Subject: Re: [Users] Disk Migration On 02/26/2014 05:24 PM, Maurice James wrote: I have a specific interface set up for migrations. Why do disk migrations not use the interface that I have set for migrations? Is that by design? Shouldnt it use the interfaces that I have set aside for migrations? VM migrations work as they should but not disk migrations I don't think that you can configure interface for disk migration. Disk migration is actually copy of information from the original disk to a new disk created on a new domain + delete of the original disk once that is done. it's not actually a migration and so I am not sure you can actually configure an interface for that. adding ofer - perhpas he has a solution or it's possible and I am not aware of it. I guess that *not* using the migration network for storage operation is the expected behavior, to make migration faster and safer. Michal, Dan, can you elaborate on this? with storage offloading it's probably not going to be significant, however today it likely is. Nir, why would not using migration network make it better? Won't we have the same problem as before without migration network at all, i.e. choking the management channel? Should we maybe consider a dedicated storage network? As far as I understand Nir, you are saying the same thing: it's better to separate management, migration, and storage to separate networks, so that one does not choke the other. When we get host network QoS done (hopefully ovirt-3.5) we can do this quite cheaply, over vlans. Dan. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Migration Failed
On Thu, Feb 27, 2014 at 10:42:54AM +0100, Koen Vanoppen wrote: Sorry... I added the correct one now Still, I fail to find the relevant ::ERROR:: line about migration. But as Michal mentioned, Could not connect to peer VDS means that source vdsm failed to contact the destination one. This can stem from physical or logical network problem. Can you ping from source to dest? Can what happens when you log into source host and run vdsClient -s fqdn-of-destination-host list ? do you get any response? What happens if you disable your firewall? Regards, Dan. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] gluster
Looks like one of your peers is not connected anymore, depending on your gluster setup this could be harmless so long as you replace it soon. On Thu, Feb 27, 2014 at 1:54 AM, yfw...@daicy.net yfw...@daicy.net wrote: hi, it is my glusterfs log, [2014-02-27 10:44:35.565367] I [client.c:1883:client_rpc_notify] 0-ctdb-client-0: disconnected [2014-02-27 10:44:35.581812] W [socket.c:1494:__socket_proto_state_machine] 0-glusterfs: reading from socket failed. Error (Transport endpoint is not connected), peer ( 192.168.101.93:24007) Is it harmless or do I need to worry? can you help me,thanks? -- yfw...@daicy.net ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Patrick Pierson ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] Fresh installation of ovirt-engine
Hello, I just finnished an installation of a ovirt-engine and changed the ISO_DOMAIN path to /exports/iso as this is my 1Tb diskspace. But now I see that the nfs share created is /var/lib/exports/iso. I would expect it to be /export/iso but that doesn't seem to be the case. Doe I need to modify this manualy or did I do something wrong. Kind regards. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning - bandwidth accounting
On Thu, Feb 27, 2014 at 02:14:32PM +0100, Johan Kooijman wrote: Dan, How about storing the rx_byte per 5 minutes in the engine DB? That way a reset of the counters has a minimal impact and analytics as traffic for VM x in month Y could be made. Another approach could be to have iptables keep the count? I do not see how iptables can help here. On the host, Vdsm reports how much traffic did VM x consume. However, when the VM is migrated to another host, the accounting at the destination are reset. If Engine simply copied the value, you can start month Y with 10GiB of traffic, and end it with 7KiB. This should be solved by Engine, probably by banking the Vdsm-reported values on certain occasions (vm shutdown, migration, counter reset), and exposing only the accumulated result. Dan. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] Snapshot merging and the effect on underlying LV metadata
Hi I am being told that unless the Wipe After Delete option is set on a vDisk, any subsequent snapshot merging of the related VM will not delete LV metadata (or any data!) from the volume created by the snapshot. Is this correct ? I'm kinda hoping not ! Richard Davis Technical Specialist PGDS Midrange (UK) - Solaris Linux This email is confidential and should not be used by anyone who is not the original intended recipient. PGDS cannot accept liability for statements made which are clearly the sender's own and not made on behalf of the company. In addition no statement should be construed as giving investment advice within or outside the United Kingdom. PGDS (UK ONE) LIMITED, Laurence Pountney Hill, London, EC4R 0HH. Incorporated and registered in England and Wales. Registered Office as above. Registered number 1967719. PGDS is the trading name of certain subsidiaries of Prudential plc (registered in England, number 1397169), whose registered office is at Laurence Pountney Hill London EC4R OHH, some of whose subsidiaries are authorised and regulated, as applicable, by the Prudential Regulation Authority and the Financial Conduct Authority. Prudential plc is not affiliated in any manner with Prudential Financial, Inc, a company whose principal place of business is in the United States of America. A list of other Prudential companies together with their registered statutory details can be found in 'About Prudential' on http://www.prudential.co.uk An e-mail reply to this address may be subject to interception or monitoring for operational reasons or for lawful business practices. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] Snapshot merging and the effect on underlying LV metadata
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 On Thu, 27 Feb 2014, Davis, Richard wrote: I am being told that unless the Wipe After Delete option is set on a vDisk, any subsequent snapshot merging of the related VM will not delete LV metadata (or any data!) from the volume created by the snapshot. Is this correct ? I'm kinda hoping not ! It is my belief a depetion cannot be relied upon to have happened in all cases. Some options flag sets in lvm ** do ** persist old data, and so our security practice at PMman to treat data on removed LV's as though it persists There are published reports that instances on other public cloud providers have been deployed with 'non-wiped' drives in the 'slack space'. Why run the reputational risk? When we reclaim a LV, we perform a 'renaming' that permits to spot 'dirty' and 'scratched' instances needing wiping. [we also fill a new VG / PV with LV's indicating it needs wiping, as we do not wish to expose content if a drive is pulled and then re-used after testing when SMART errors appeared, but do not stand up to disqualify a drive] Later a cron driven process, sensitive to IO load runs. It builds a list of candidates over a day old, using 'find' and the LV name series showing it is dirty and scratched. Then in turn by LV found, it fires off a sub-task (when load is low), which in turn performs a 'niced' 'shred' operation on that LV, followed by the 'shred 'zeroing' operation. When load is too high, it sleeps for a couple of minutes, and re-tries fragment: $_shredCmd = ionice -c 3 shred -n \ .$_num_passes. -z .$_working_lvm; Only when that sub-process has completed do we 'rename' and later 'remove' a given LV, to let its space re-enter the assignment pool - -- Russ herrold -BEGIN PGP SIGNATURE- Version: GnuPG v2.0.14 (GNU/Linux) iEYEARECAAYFAlMPkAMACgkQMRh1QZtklkSamQCgnVqEo2Kmzq9Ao8T0BCYhBTyn aToAoIaOVGkxX3EsVghMxOtgE3RiUr9G =rm/K -END PGP SIGNATURE- ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
I think it will be nice to have: 1. A VM appliance to do backups using the backup APIs or add this functionality in engine 2. engine to create and maintain gluster geo-replication 3. freeBSD as supported guest and guest agent for it Regards Liviu On Fri, Feb 28, 2014 at 2:42 AM, Maurice James midnightst...@msn.comwrote: Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com To: jo...@israelsson.com; users@ovirt.org; mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users inline: poweroff.PNG___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
On 02/27/2014 09:35 PM, Liviu Elama wrote: I think it will be nice to have: 1. A VM appliance to do backups using the backup APIs or add this functionality in engine I'd rather see integrated with backup solutions to tackle this (hopefully, there are relevant open source ones as well) 2. engine to create and maintain gluster geo-replication sahina - thoughts? 3. freeBSD as supported guest and guest agent for it adam - where are we with getting freebsd added as guest? http://gerrit.ovirt.org/#/c/22273/ Liviu - note it is just a config file (other than the icon). vinzenz - i don't remember anyone volunteered to test and package the guest agent on freebsd? (liviu - maybe you can try to tackle that?) Regards Liviu On Fri, Feb 28, 2014 at 2:42 AM, Maurice James midnightst...@msn.com mailto:midnightst...@msn.com wrote: Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com mailto:ih...@redhat.com To: jo...@israelsson.com mailto:jo...@israelsson.com; users@ovirt.org mailto:users@ovirt.org; mskri...@redhat.com mailto:mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
On 27/02/14 21:44 +0200, Itamar Heim wrote: On 02/27/2014 09:35 PM, Liviu Elama wrote: I think it will be nice to have: 1. A VM appliance to do backups using the backup APIs or add this functionality in engine I'd rather see integrated with backup solutions to tackle this (hopefully, there are relevant open source ones as well) 2. engine to create and maintain gluster geo-replication sahina - thoughts? 3. freeBSD as supported guest and guest agent for it adam - where are we with getting freebsd added as guest? http://gerrit.ovirt.org/#/c/22273/ The patch is Code Reviewed and Verified. Just needs to be pushed. Liviu - note it is just a config file (other than the icon). vinzenz - i don't remember anyone volunteered to test and package the guest agent on freebsd? (liviu - maybe you can try to tackle that?) Regards Liviu On Fri, Feb 28, 2014 at 2:42 AM, Maurice James midnightst...@msn.com mailto:midnightst...@msn.com wrote: Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com mailto:ih...@redhat.com To: jo...@israelsson.com mailto:jo...@israelsson.com; users@ovirt.org mailto:users@ovirt.org; mskri...@redhat.com mailto:mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Adam Litke ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
1. A VM appliance to do backups using the backup APIs or add this functionality in engine I'd rather see integrated with backup solutions to tackle this (hopefully, there are relevant open source ones as well) Do you know if there is a 3rd party that has started work on this? 2. engine to create and maintain gluster geo-replication sahina - thoughts? 3. freeBSD as supported guest and guest agent for it adam - where are we with getting freebsd added as guest? http://gerrit.ovirt.org/#/c/22273/ Liviu - note it is just a config file (other than the icon). vinzenz - i don't remember anyone volunteered to test and package the guest agent on freebsd? (liviu - maybe you can try to tackle that?) I can surely test the agent and I can try give it a go to package agent for 8.3(if no one else did this before) as I'm interested in pfsense on ovirt Regards Liviu On Fri, Feb 28, 2014 at 2:42 AM, Maurice James midnightst...@msn.com mailto:midnightst...@msn.com wrote: Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com mailto:ih...@redhat.com To: jo...@israelsson.com mailto:jo...@israelsson.com; users@ovirt.org mailto:users@ovirt.org; mskri...@redhat.com mailto:mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
On 02/27/2014 10:51 PM, Liviu Elama wrote: 1. A VM appliance to do backups using the backup APIs or add this functionality in engine I'd rather see integrated with backup solutions to tackle this (hopefully, there are relevant open source ones as well) Do you know if there is a 3rd party that has started work on this? I'm aware of several, yes. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] The purpose of Wipe on delete ?
Hi What is the purpose of the Wipe on delete option for a VM disk ? Why would you not want data wiped on delete if the alternative is to leave LV metadata and other data languishing on the SD ? Thanks Rich ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] The purpose of Wipe on delete ?
wipe = writing zero's on the space allocated to that disk to make sure any data once written will be deleted permanently. so it's a security vs speed decision on using this option - since we zeroing the disk to make sure any information once written will be overwritten, a delete of a large disk can take a while. Dafna On 02/27/2014 04:14 PM, Richard Davis wrote: Hi What is the purpose of the Wipe on delete option for a VM disk ? Why would you not want data wiped on delete if the alternative is to leave LV metadata and other data languishing on the SD ? Thanks Rich ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Dafna Ron ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Fresh installation of ovirt-engine
On 02/27/2014 06:14 PM, Andy Michielsen wrote: Hello, I just finnished an installation of a ovirt-engine and changed the ISO_DOMAIN path to /exports/iso as this is my 1Tb diskspace. But now I see that the nfs share created is /var/lib/exports/iso. I would expect it to be /export/iso but that doesn't seem to be the case. Doe I need to modify this manualy or did I do something wrong. Kind regards. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users didi/sandro - thoughts? ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
Den 27 feb 2014 21:51 skrev Liviu Elama liviu.el...@gmail.com: 1. A VM appliance to do backups using the backup APIs or add this functionality in engine I'd rather see integrated with backup solutions to tackle this (hopefully, there are relevant open source ones as well) Do you know if there is a 3rd party that has started work on this? 2. engine to create and maintain gluster geo-replication sahina - thoughts? 3. freeBSD as supported guest and guest agent for it adam - where are we with getting freebsd added as guest? http://gerrit.ovirt.org/#/c/22273/ Liviu - note it is just a config file (other than the icon). vinzenz - i don't remember anyone volunteered to test and package the guest agent on freebsd? (liviu - maybe you can try to tackle that?) I can surely test the agent and I can try give it a go to package agent for 8.3(if no one else did this before) as I'm interested in pfsense on ovirt Don't, it's a moot point until someone gets down to sorting out virtio-serial: http://lists.ovirt.org/pipermail/users/2013-August/015775.html This is the most recent info I could find on the matter: http://lists.freebsd.org/pipermail/freebsd-emulation/2013-April/010569.html You should get in touch with that guy, let him know there's interest at least:) /K Regards Liviu On Fri, Feb 28, 2014 at 2:42 AM, Maurice James midnightst...@msn.com mailto:midnightst...@msn.com wrote: Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com mailto:ih...@redhat.com To: jo...@israelsson.com mailto:jo...@israelsson.com; users@ovirt.org mailto:users@ovirt.org; mskri...@redhat.com mailto:mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] The purpose of Wipe on delete ?
On Thu, 27 Feb 2014, Richard Davis wrote: What is the purpose of the Wipe on delete option for a VM disk ? Why would you not want data wiped on delete if the alternative is to leave LV metadata and other data languishing on the SD ? debugging, cloning, snapshot backups, and giving 'second chances' to undelete, come to mind as reasonable use cases NOT to wipe by default -- Russ herrold ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[Users] Notifier not sending as correct user
I set up the engine notifier to send notifications, but when the email is delivered they are coming from ovirt@hostname.domain instead of the address that I have specified in the MAIL_USER and MAIL_FROM section of the ovirt-engine-notifier.conf file ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Notifier not sending as correct user
I forgot to add that im using version 3.4.0-0.9.beta3.el6 From: midnightst...@msn.com To: users@ovirt.org Date: Thu, 27 Feb 2014 21:33:35 -0500 Subject: [Users] Notifier not sending as correct user I set up the engine notifier to send notifications, but when the email is delivered they are coming from ovirt@hostname.domain instead of the address that I have specified in the MAIL_USER and MAIL_FROM section of the ovirt-engine-notifier.conf file ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] oVirt 3.5 planning
On 02/27/2014 09:51 PM, Liviu Elama wrote: 1. A VM appliance to do backups using the backup APIs or add this functionality in engine I'd rather see integrated with backup solutions to tackle this (hopefully, there are relevant open source ones as well) Do you know if there is a 3rd party that has started work on this? 2. engine to create and maintain gluster geo-replication sahina - thoughts? 3. freeBSD as supported guest and guest agent for it adam - where are we with getting freebsd added as guest? http://gerrit.ovirt.org/#/c/22273/ Liviu - note it is just a config file (other than the icon). vinzenz - i don't remember anyone volunteered to test and package the guest agent on freebsd? The latest information I have is, that they yet don't have virtio serial drivers which is the very basic requirement. (liviu - maybe you can try to tackle that?) I can surely test the agent and I can try give it a go to package agent for 8.3(if no one else did this before) as I'm interested in pfsense on ovirt Regards Liviu On Fri, Feb 28, 2014 at 2:42 AM, Maurice James midnightst...@msn.com mailto:midnightst...@msn.com mailto:midnightst...@msn.com mailto:midnightst...@msn.com wrote: Its a plug icon Date: Thu, 27 Feb 2014 15:31:34 +0200 From: ih...@redhat.com mailto:ih...@redhat.com mailto:ih...@redhat.com mailto:ih...@redhat.com To: jo...@israelsson.com mailto:jo...@israelsson.com mailto:jo...@israelsson.com mailto:jo...@israelsson.com; users@ovirt.org mailto:users@ovirt.org mailto:users@ovirt.org mailto:users@ovirt.org; mskri...@redhat.com mailto:mskri...@redhat.com mailto:mskri...@redhat.com mailto:mskri...@redhat.com Subject: Re: [Users] oVirt 3.5 planning On 02/25/2014 12:00 PM, Jonas Israelsson wrote: Not sure if this already exist but I have had to help quite a few users that have only user-permissions to restart their VM if they managed to hang the OS. This since they lack the permission to power off the machine, and shutdown simply is not enough. Giving them more permission can help, since they then will have the extended tab with more options including the ability to power off a VM , this however IMO is overkill since they are then presented with a vast number of options such as add disk, nic, networks etc, all not working since they have no (and should have none) permission to those objects. So adding to the basic view in the user portal a power off button and extending the ordinary user-permission to also include power off I think would be a good idea. Rgds Jonas ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org mailto:Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users michal - don't we have power off vm in the basic user portal? ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org mailto:Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org mailto:Users@ovirt.org mailto:Users@ovirt.org mailto:Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users -- Regards, Vinzenz Feenstra | Senior Software Engineer RedHat Engineering Virtualization R D Phone: +420 532 294 625 IRC: vfeenstr or evilissimo Better technology. Faster innovation. Powered by community collaboration. See how it works at redhat.com ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [Users] Fresh installation of ovirt-engine
Il 27/02/2014 22:18, Itamar Heim ha scritto: On 02/27/2014 06:14 PM, Andy Michielsen wrote: Hello, I just finnished an installation of a ovirt-engine and changed the ISO_DOMAIN path to /exports/iso as this is my 1Tb diskspace. But now I see that the nfs share created is /var/lib/exports/iso. I would expect it to be /export/iso but that doesn't seem to be the case. Doe I need to modify this manualy or did I do something wrong. So you used default /var/lib/exports/iso while running engine setup and then changed it manually from the web app? Can you upload your setup logs (for example on http://www.fpaste.org/) so we can take a look? Kind regards. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users didi/sandro - thoughts? -- Sandro Bonazzola Better technology. Faster innovation. Powered by community collaboration. See how it works at redhat.com ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users