On Tue, Feb 21, 2017 at 7:25 PM, Gianluca Cecchi <[email protected]> wrote: > On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer <[email protected]> wrote: >> >> This is caused by active lvs on the remove storage domains that were not >> deactivated during the removal. This is a very old known issue. >> >> You have remove the remove device mapper entries - you can see the devices >> using: >> >> dmsetup status >> >> Then you can remove the mapping using: >> >> dmsetup remove device-name >> >> Once you removed the stale lvs, you will be able to remove the multipath >> device and the underlying paths, and lvm will not complain about read >> errors. >> >> Nir > > > OK Nir, thanks for advising. > > So what I run with success on the 2 hosts > > [root@ovmsrv05 vdsm]# for dev in $(dmsetup status | grep > 900b1853--e192--4661--a0f9--7c7c396f6f49 | cut -d ":" -f 1) > do > dmsetup remove $dev > done > [root@ovmsrv05 vdsm]# > > and now I can run > > [root@ovmsrv05 vdsm]# multipath -f 3600a0b80002999020000cd3c5501458f > [root@ovmsrv05 vdsm]# > > Also, with related names depending on host, > > previous maps to single devices were for example in ovmsrv05: > > 3600a0b80002999020000cd3c5501458f dm-4 IBM ,1814 FAStT > size=2.0T features='2 pg_init_retries 50' hwhandler='1 rdac' wp=rw > |-+- policy='service-time 0' prio=0 status=enabled > | |- 0:0:0:2 sdb 8:16 failed undef running > | `- 1:0:0:2 sdh 8:112 failed undef running > `-+- policy='service-time 0' prio=0 status=enabled > |- 0:0:1:2 sdg 8:96 failed undef running > `- 1:0:1:2 sdn 8:208 failed undef running > > And removal of single path devices: > > [root@ovmsrv05 root]# for dev in sdb sdh sdg sdn > do > echo 1 > /sys/block/${dev}/device/delete > done > [root@ovmsrv05 vdsm]# > > All clean now... ;-)
Great! I think we should have a script doing all these steps. Nir _______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

