Hi Nir, Thanks for solution. I didn't notice the guest /dev/backupvg01/backuplv01 on all hypervisors. It seems that I've got this issue with 2 additionals volumes, but no one noticed because they were only few gb.
[root@wrops2 BLUE/WRO ~]# ls -l /sys/block/$(basename $(readlink /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8))/holders total 0 lrwxrwxrwx. 1 root root 0 Jun 13 10:48 dm-43 -> ../../dm-43 [root@wrops2 BLUE/WRO ~]# pvscan --cache [root@wrops2 BLUE/WRO ~]# vgs -o pv_name,vg_name PV VG /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 backupvg01 /dev/sda2 centos_wrops2 /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/99a1c067-9728-484a-a0cb-cb6689d5724c deployvg /dev/mapper/3600000e00d0000000024057200000000 e69d1c16-36d1-4375-aaee-69f5a5ce1616 /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/86a6d83f-2661-4fe3-8874-ce4d8a111c0d jenkins /dev/sda3 w2vg1 [root@wrops2 BLUE/WRO ~]# dmsetup info Name: backupvg01-backuplv01 State: ACTIVE Read Ahead: 8192 Tables present: LIVE Open count: 0 Event number: 0 Major, minor: 253, 43 Number of targets: 1 UUID: LVM-ubxOH5R2h6B8JwLGfhpiNjnAKlPxMPy6KfkeLBxXajoT3gxU0yC5JvOQQVkixrTA [root@wrops2 BLUE/WRO ~]# lvchange -an /dev/backupvg01/backuplv01 [root@wrops2 BLUE/WRO ~]# lvremove /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 Do you really want to remove active logical volume ee53af81-820d-4916-b766-5236ca99daf8? [y/n]: y Logical volume "ee53af81-820d-4916-b766-5236ca99daf8" successfully removed Would this configuration in lvm.conf: filter = [ "r|/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/.*|" ] on all hypervisors solve problem of scanning guest volumes? 2016-06-11 23:16 GMT+02:00 Nir Soffer <nsof...@redhat.com>: > On Thu, Jun 9, 2016 at 11:46 AM, Krzysztof Dajka <alter...@gmail.com> > wrote: > > Hi, > > > > Recently I tried to delete 1TB disk created on top ~3TB LUN from > > ovirtengine. > > Disk is preallocated and I backuped data to other disk so I could > recreate > > it once again as thin volume. I couldn't remove this disk when it was > > attached to a VM. But once I detached it I could remove it permanently. > The > > thing is it only disappeared from ovirtengine GUI. > > > > I've got 4 hosts with FC HBA attached to storage array and all of them > are > > saying that this 1TB disk which should be gone is opened by all hosts > > simultaneously. > > > > [root@wrops1 BLUE ~]# lvdisplay -m > > > /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 > > --- Logical volume --- > > LV Path > > > /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 > > LV Name ee53af81-820d-4916-b766-5236ca99daf8 > > VG Name e69d1c16-36d1-4375-aaee-69f5a5ce1616 > > LV UUID sBdBRk-tNyZ-Rval-F4lw-ka6X-wOe8-AQenTb > > LV Write Access read/write > > LV Creation host, time wrops1.blue, 2015-07-31 10:40:57 +0200 > > LV Status available > > # open 1 > > LV Size 1.00 TiB > > Current LE 8192 > > Segments 1 > > Allocation inherit > > Read ahead sectors auto > > - currently set to 8192 > > Block device 253:29 > > > > --- Segments --- > > Logical extents 0 to 8191: > > Type linear > > Physical volume /dev/mapper/3600000e00d0000000024057200000000 > > Physical extents 8145 to 16336 > > > > Deactivating LV doesn't work: > > [root@wrops1 BLUE ~]# lvchange -an > > > /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 > > Logical volume > > > e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 is > > used by another device. > > Looks like your lv is used as a physical volume on another vg - probably > a vg created on a guest. Lvm and systemd are trying hard to discover > stuff on multipath devices and expose anything to the hypervisor. > > Can you share the output of: > > ls -l /sys/block/$(basename $(readlink > > /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8))/holders > > And: > > pvscan --cache > vgs -o pv_name,vg_name > > Nir > > > Removing from hypervisor doesn't work either. > > [root@wrops1 BLUE ~]# lvremove --force > > > /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 > > Logical volume > > > e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 is > > used by another device. > > > > I tried and rebooted one host and as soon as it booted the volume became > > opened once again. Lsof on all hosts doesn't give anything meaningful > > regarding this LV. As opposed to other LV which are used by qemu-kvm. > > > > Has anyone encountered similar problem? How can I remove this LV? > > > > _______________________________________________ > > Users mailing list > > Users@ovirt.org > > http://lists.ovirt.org/mailman/listinfo/users > > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users