cat /proc/mdstat, the 253,19 is likely an /dev/mdX device and to get an io error like that it has to be in a wrong state.
On Mon, Mar 23, 2020 at 5:14 AM Bernd Eckenfels <e...@zusammenkunft.net> wrote: > > Do you see any dmesg kernel errors when you try to activate the LVs? > > Gruss > Bernd > > > -- > http://bernd.eckenfels.net > ________________________________ > Von: linux-lvm-boun...@redhat.com <linux-lvm-boun...@redhat.com> im Auftrag > von Andrew Falgout <digitalw...@gmail.com> > Gesendet: Saturday, March 21, 2020 4:22:04 AM > An: linux-lvm@redhat.com <linux-lvm@redhat.com> > Betreff: [linux-lvm] lvm raid5 : drives all present but vg/lvm will not > assemble > > > This started on a Raspberry PI 4 running raspbian. I moved the disks to my > Fedora 31 system, that is running the latest updates and kernel. When I had > the same issues there I knew it wasn't raspbian. > > I've reached the end of my rope on this. The disks are there, all three are > accounted for, and the LVM data on them can be seen. But it refuses to > activate stating I/O errors. > > [root@hypervisor01 ~]# pvs > PV VG Fmt Attr PSize PFree > /dev/sda1 local_storage01 lvm2 a-- <931.51g 0 > /dev/sdb1 local_storage01 lvm2 a-- <931.51g 0 > /dev/sdc1 local_storage01 lvm2 a-- <931.51g 0 > /dev/sdd1 local_storage01 lvm2 a-- <931.51g 0 > /dev/sde1 local_storage01 lvm2 a-- <931.51g 0 > /dev/sdf1 local_storage01 lvm2 a-- <931.51g <931.51g > /dev/sdg1 local_storage01 lvm2 a-- <931.51g <931.51g > /dev/sdh1 local_storage01 lvm2 a-- <931.51g <931.51g > /dev/sdi3 fedora_hypervisor lvm2 a-- 27.33g <9.44g > /dev/sdk1 vg1 lvm2 a-- <7.28t 0 > /dev/sdl1 vg1 lvm2 a-- <7.28t 0 > /dev/sdm1 vg1 lvm2 a-- <7.28t 0 > [root@hypervisor01 ~]# vgs > VG #PV #LV #SN Attr VSize VFree > fedora_hypervisor 1 2 0 wz--n- 27.33g <9.44g > local_storage01 8 1 0 wz--n- <7.28t <2.73t > vg1 3 1 0 wz--n- 21.83t 0 > [root@hypervisor01 ~]# lvs > LV VG Attr LSize Pool Origin Data% Meta% > Move Log Cpy%Sync Convert > root fedora_hypervisor -wi-ao---- 15.00g > swap fedora_hypervisor -wi-ao---- 2.89g > libvirt local_storage01 rwi-aor--- <2.73t > 100.00 > gluster02 vg1 Rwi---r--- 14.55t > > The one in question is the vg1/gluster02 lvm group. > > I try to activate the VG: > [root@hypervisor01 ~]# vgchange -ay vg1 > device-mapper: reload ioctl on (253:19) failed: Input/output error > 0 logical volume(s) in volume group "vg1" now active > > I've got the debugging output from : > vgchange -ay vg1 -vvvv -dddd > lvchange -ay --partial vg1/gluster02 -vvvv -dddd > > Just not sure where I should dump the data for people to look at. Is there a > way to tell the md system to ignore the metadata since there wasn't an actual > disk failure, and rebuild the metadata off what is in the lvm? Or can I even > get the LV to mount, so I can pull the data off. > > Any help is appreciated. If I can save the data great. I'm tossing this to > the community to see if anyone else has an idea of what I can do. > ./digitalw00t > _______________________________________________ > linux-lvm mailing list > linux-lvm@redhat.com > https://www.redhat.com/mailman/listinfo/linux-lvm > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ _______________________________________________ linux-lvm mailing list linux-lvm@redhat.com https://www.redhat.com/mailman/listinfo/linux-lvm read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/