I can create snapshot when no one exists but I'm not able to remove it after. It concerns many of my vms, and when stopping them, they can't boot anymore because of the illegal status of the disks, this leads me in a critical situation

VM fedora23 is down with error. Exit message: Unable to get volume size for domain 5ef8572c-0ab5-4491-994a-e4c30230a525 volume e5969faa-97ea-41df-809b-cc62161ab1bc

As far as I didn't initiate any live merge, am I concerned by this bug https://bugzilla.redhat.com/show_bug.cgi?id=1306741?
I'm running 3.6.2, will upgrade to 3.6.3 solve this issue?

2016-03-18 18:26:57,652 ERROR [org.ovirt.engine.core.bll.RemoveSnapshotCommand] (org.ovirt.thread.pool-8-thread-39) [a1e222d] Ending command 'org.ovirt.engine.core.bll.RemoveSnapshotCommand' with failure. 2016-03-18 18:26:57,663 ERROR [org.ovirt.engine.core.bll.RemoveSnapshotCommand] (org.ovirt.thread.pool-8-thread-39) [a1e222d] Could not delete image '46e9ecc8-e168-4f4d-926c-e769f5df1f2c' from snapshot '88fcf167-4302-405e-825f-ad7e0e9f6564' 2016-03-18 18:26:57,678 WARN [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (org.ovirt.thread.pool-8-thread-39) [a1e222d] Correlation ID: a1e222d, Job ID: 00d3e364-7e47-4022-82ff-f772cd79d4a1, Call Stack: null, Custom Event ID: -1, Message: Due to partial snapshot removal, Snapshot 'test' of VM 'fedora23' now contains only the following disks: 'fedora23_Disk1'. 2016-03-18 18:26:57,695 ERROR [org.ovirt.engine.core.bll.RemoveSnapshotSingleDiskCommand] (org.ovirt.thread.pool-8-thread-39) [724e99fd] Ending command 'org.ovirt.engine.core.bll.RemoveSnapshotSingleDiskCommand' with failure. 2016-03-18 18:26:57,708 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandlin

Thank you for your help.

Le 23/02/2016 19:51, Greg Padgett a écrit :
On 02/22/2016 07:10 AM, Marcelo Leandro wrote:

The bug with snapshot  it will be fixed in ovirt 3.6.3?


Hi Marcelo,

Yes, the bug below (bug 1301709) is now targeted to 3.6.3.


2016-02-18 11:34 GMT-03:00 Adam Litke <ali...@redhat.com>:
On 18/02/16 10:37 +0100, Rik Theys wrote:


On 02/17/2016 05:29 PM, Adam Litke wrote:

On 17/02/16 11:14 -0500, Greg Padgett wrote:

On 02/17/2016 03:42 AM, Rik Theys wrote:


On 02/16/2016 10:52 PM, Greg Padgett wrote:

On 02/16/2016 08:50 AM, Rik Theys wrote:

  From the above I conclude that the disk with id that ends with

Similar to what I wrote to Marcelo above in the thread, I'd recommend running the "VM disk info gathering tool" attached to [1]. It's the best way to ensure the merge was completed and determine which image
the "bad" one that is no longer in use by any volume chains.

I've ran the disk info gathering tool and this outputs (for the

VM lena
     Disk b2390535-744f-4c02-bdc8-5a897226554b

The id of the volume is the ID of the snapshot that is marked
So the "bad" image would be the dc39 one, which according to the UI is
in use by the "Active VM" snapshot. Can this make sense?

It looks accurate.  Live merges are "backwards" merges, so the merge
would have pushed data from the volume associated with "Active VM"
into the volume associated with the snapshot you're trying to remove.

Upon completion, we "pivot" so that the VM uses that older volume, and
we update the engine database to reflect this (basically we
re-associate that older volume with, in your case, "Active VM").

In your case, it seems the pivot operation was done, but the database wasn't updated to reflect it. Given snapshot/image associations e.g.:

  VM Name  Snapshot Name  Volume
  -------  -------------  ------
  My-VM    Active VM      123-abc
  My-VM    My-Snapshot    789-def

My-VM in your case is actually running on volume 789-def. If you run
the db fixup script and supply ("My-VM", "My-Snapshot", "123-abc")
(note the volume is the newer, "bad" one), then it will switch the
volume association for you and remove the invalid entries.

Of course, I'd shut down the VM, and back up the db beforehand.

I've executed the sql script and it seems to have worked. Thanks!

"Active VM" should now be unused; it previously (pre-merge) was the
data written since the snapshot was taken. Normally the larger actual
size might be from qcow format overhead.  If your listing above is
complete (ie one volume for the vm), then I'm not sure why the base
volume would have a larger actual size than virtual size.

Adam, Nir--any thoughts on this?

There is a bug which has caused inflation of the snapshot volumes when
performing a live merge.  We are submitting fixes for 3.5, 3.6, and
master right at this moment.

Which bug number is assigned to this bug? Will upgrading to a release
with a fix reduce the disk usage again?

See https://bugzilla.redhat.com/show_bug.cgi?id=1301709 for the bug.
It's about a clone disk failure after the problem occurs.
Unfortunately, there is not an automatic way to repair the raw base
volumes if they were affected by this bug.  They will need to be
manually shrunk using lvreduce if you are certain that they are

Adam Litke

Users mailing list
Users mailing list

Users mailing list

Nathanaël Blanchet

Supervision réseau
Pôle Infrastrutures Informatiques
227 avenue Professeur-Jean-Louis-Viala
34193 MONTPELLIER CEDEX 5       
Tél. 33 (0)4 67 54 84 55
Fax  33 (0)4 67 54 84 14

Users mailing list

Reply via email to