------- Comment From [email protected] 2017-02-22 00:29 EDT-------
(In reply to comment #30)
>
> Hi,
> as usual bugproxy accumulated a lot of info to go through.
> I'll try to summarize - please feel free to correct - "VG/LV are not
> available in /dev/mapper/ after reboot".
> Would that be a proper (and simpler) definition of the issue?
>
> I tried to recreate.
> Note: as it was reported on power + Xenial this test was done on ppc64el
> Xenial as of today
>
> To do so with as much debugging as possible I created a normal Xenial KVM
> Guest via
> $ uvt-kvm create --cpu 4 --password=ubuntu paelzer-testlvm-xenial
> release=xenial
> Then I added a few more disks to be used as PVs
> $ sudo qemu-img create -f qcow2 test-lvm-disk1.qcow2 8
> And added those to the Guest.
>
> The guest then initially looks like:
> $ lsblk
> NAME   MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
> vda    253:0    0    8G  0 disk
> vdb    253:16   0    8G  0 disk
> vdc    253:32   0    8G  0 disk
> vdd    253:48   0  366K  0 disk
> vde    253:64   0    8G  0 disk
> |-vde1 253:65   0    8G  0 part /
> `-vde2 253:66   0    8M  0 part
>
> Then the usual flow is
> 1. fdisk, create partition set LVM partition type (8e)
> $ sudo fdisk /dev/vd[abc]
> 2. Full PVs on all the three disks
> $ sudo pvcreate /dev/vd[abc]
> 3. vgcreate a single VG out of all of the PVs
> $ sudo vgcreate vg /dev/vda1 /dev/vdb1 /dev/vdc1
>
> At this point it looks like this:
> $ sudo vgdisplay vg
> --- Volume group ---
> VG Name               vg
> System ID
> Format                lvm2
> Metadata Areas        3
> Metadata Sequence No  1
> VG Access             read/write
> VG Status             resizable
> MAX LV                0
> Cur LV                0
> Open LV               0
> Max PV                0
> Cur PV                3
> Act PV                3
> VG Size               23.99 GiB
> PE Size               4.00 MiB
> Total PE              6141
> Alloc PE / Size       0 / 0
> Free  PE / Size       6141 / 23.99 GiB
> VG UUID               1QMFbn-5DAW-T9IE-Fdfd-9RZK-t8gl-5nte5r
>
> Ok, create normal as well as thin LVs out of that now.
> First of all thin provisioning is not mainstream, the dependency is only a
> suggest, so install the tools
> $ sudo apt-get install thin-provisioning-tools
> Then create the normal LV
> $ sudo lvcreate -L 5G --name lv_normal vg
> And finally a thin LV
> $ sudo lvcreate --size 10G --virtualsize 5G --thinpool mythinpool --name
> lv_thin vg
> Lets go harder and overprovision the thinpool
> $ sudo lvcreate --virtualsize 5G --thinpool mythinpool --name lv_thin2 vg
> $ sudo lvcreate --virtualsize 5G --thinpool mythinpool --name lv_thin3 vg
>
> With that in place my LVs look like:
> $ sudo lvdisplay
> --- Logical volume ---
> LV Path                /dev/vg/lv_normal
> LV Name                lv_normal
> VG Name                vg
> LV UUID                aCtNC0-gbx1-uHoB-3dC8-dfhl-NBxd-Axm879
> LV Write Access        read/write
> LV Creation host, time paelzer-testlvm-xenial, 2017-01-25 12:54:02 +0000
> LV Status              available
> # open                 0
> LV Size                5.00 GiB
> Current LE             1280
> Segments               1
> Allocation             inherit
> Read ahead sectors     auto
> - currently set to     256
> Block device           252:0
>
> --- Logical volume ---
> LV Name                mythinpool
> VG Name                vg
> LV UUID                UCKy8A-ovc6-Qh9n-wrEM-c2s2-myvD-hpLhb0
> LV Write Access        read/write
> LV Creation host, time paelzer-testlvm-xenial, 2017-01-25 13:04:03 +0000
> LV Pool metadata       mythinpool_tmeta
> LV Pool data           mythinpool_tdata
> LV Status              available
> # open                 4
> LV Size                10.00 GiB
> Allocated pool data    0.00%
> Allocated metadata     0.72%
> Current LE             2560
> Segments               1
> Allocation             inherit
> Read ahead sectors     auto
> - currently set to     256
> Block device           252:3
>
> --- Logical volume ---
> LV Path                /dev/vg/lv_thin
> LV Name                lv_thin
> VG Name                vg
> LV UUID                ay7Clp-78K7-8UoZ-ueZZ-WYTy-IQlF-oQkjdT
> LV Write Access        read/write
> LV Creation host, time paelzer-testlvm-xenial, 2017-01-25 13:04:03 +0000
> LV Pool name           mythinpool
> LV Status              available
> # open                 0
> LV Size                5.00 GiB
> Mapped size            0.00%
> Current LE             1280
> Segments               1
> Allocation             inherit
> Read ahead sectors     auto
> - currently set to     256
> Block device           252:5
>
> --- Logical volume ---
> LV Path                /dev/vg/lv_thin2
> LV Name                lv_thin2
> VG Name                vg
> LV UUID                Zj5BfS-Cbm1-9TgI-gwk3-vdBG-wwH9-cSmC0B
> LV Write Access        read/write
> LV Creation host, time paelzer-testlvm-xenial, 2017-01-25 13:05:11 +0000
> LV Pool name           mythinpool
> LV Status              available
> # open                 0
> LV Size                5.00 GiB
> Mapped size            0.00%
> Current LE             1280
> Segments               1
> Allocation             inherit
> Read ahead sectors     auto
> - currently set to     256
> Block device           252:6
>
> --- Logical volume ---
> LV Path                /dev/vg/lv_thin3
> LV Name                lv_thin3
> VG Name                vg
> LV UUID                bSPGZH-Thwx-M8qx-QDS2-gIUz-50us-hGCRGu
> LV Write Access        read/write
> LV Creation host, time paelzer-testlvm-xenial, 2017-01-25 13:05:16 +0000
> LV Pool name           mythinpool
> LV Status              available
> # open                 0
> LV Size                5.00 GiB
> Mapped size            0.00%
> Current LE             1280
> Segments               1
> Allocation             inherit
> Read ahead sectors     auto
> - currently set to     256
> Block device           252:7
>
> $ ll /dev/mapper/
> total 0
> drwxr-xr-x  2 root root     220 Jan 25 13:05 ./
> drwxr-xr-x 16 root root    4240 Jan 25 13:05 ../
> crw-------  1 root root 10, 236 Jan 25 12:20 control
> lrwxrwxrwx  1 root root       7 Jan 25 12:54 vg-lv_normal -> ../dm-0
> lrwxrwxrwx  1 root root       7 Jan 25 13:04 vg-lv_thin -> ../dm-5
> lrwxrwxrwx  1 root root       7 Jan 25 13:05 vg-lv_thin2 -> ../dm-6
> lrwxrwxrwx  1 root root       7 Jan 25 13:05 vg-lv_thin3 -> ../dm-7
> lrwxrwxrwx  1 root root       7 Jan 25 13:04 vg-mythinpool -> ../dm-4
> lrwxrwxrwx  1 root root       7 Jan 25 13:04 vg-mythinpool-tpool -> ../dm-3
> lrwxrwxrwx  1 root root       7 Jan 25 13:04 vg-mythinpool_tdata -> ../dm-2
> lrwxrwxrwx  1 root root       7 Jan 25 13:04 vg-mythinpool_tmeta -> ../dm-1
>
> Now lets put it to use
> for dev in lv_normal lv_thin lv_thin2 lv_thin3;
> do
> sudo mkfs.ext4 /dev/mapper/vg-${dev}
> sudo mkdir -p /mnt/${dev}
> sudo mount /dev/mapper/vg-${dev} /mnt/${dev}
> sudo touch /mnt/${dev}/foobar
> done
>
> Things are properly mounted
> $ mount | grep lv
> /dev/mapper/vg-lv_normal on /mnt/lv_normal type ext4
> (rw,relatime,data=ordered)
> /dev/mapper/vg-lv_thin on /mnt/lv_thin type ext4
> (rw,relatime,stripe=16,data=ordered)
> /dev/mapper/vg-lv_thin2 on /mnt/lv_thin2 type ext4
> (rw,relatime,stripe=16,data=ordered)
> /dev/mapper/vg-lv_thin3 on /mnt/lv_thin3 type ext4
> (rw,relatime,stripe=16,data=ordered)
>
> All files written normally
> $ ls -laF /mnt/*/foobar
> -rw-r--r-- 1 root root 0 Jan 25 13:10 /mnt/lv_normal/foobar
> -rw-r--r-- 1 root root 0 Jan 25 13:10 /mnt/lv_thin/foobar
> -rw-r--r-- 1 root root 0 Jan 25 13:10 /mnt/lv_thin2/foobar
> -rw-r--r-- 1 root root 0 Jan 25 13:10 /mnt/lv_thin3/foobar
>
> lsblk holds the thin LVs as it should
> $ lsblk
> NAME                      MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
> vda                       253:0    0    8G  0 disk
> `-vda1                    253:1    0    8G  0 part
> |-vg-lv_normal          252:0    0    5G  0 lvm  /mnt/lv_normal
> |-vg-mythinpool_tmeta   252:1    0   12M  0 lvm
> | `-vg-mythinpool-tpool 252:3    0   10G  0 lvm
> |   |-vg-mythinpool     252:4    0   10G  0 lvm
> |   |-vg-lv_thin        252:5    0    5G  0 lvm  /mnt/lv_thin
> |   |-vg-lv_thin2       252:6    0    5G  0 lvm  /mnt/lv_thin2
> |   `-vg-lv_thin3       252:7    0    5G  0 lvm  /mnt/lv_thin3
> `-vg-mythinpool_tdata   252:2    0   10G  0 lvm
> `-vg-mythinpool-tpool 252:3    0   10G  0 lvm
> |-vg-mythinpool     252:4    0   10G  0 lvm
> |-vg-lv_thin        252:5    0    5G  0 lvm  /mnt/lv_thin
> |-vg-lv_thin2       252:6    0    5G  0 lvm  /mnt/lv_thin2
> `-vg-lv_thin3       252:7    0    5G  0 lvm  /mnt/lv_thin3
> vdb                       253:16   0    8G  0 disk
> `-vdb1                    253:17   0    8G  0 part
> `-vg-mythinpool_tdata   252:2    0   10G  0 lvm
> `-vg-mythinpool-tpool 252:3    0   10G  0 lvm
> |-vg-mythinpool     252:4    0   10G  0 lvm
> |-vg-lv_thin        252:5    0    5G  0 lvm  /mnt/lv_thin
> |-vg-lv_thin2       252:6    0    5G  0 lvm  /mnt/lv_thin2
> `-vg-lv_thin3       252:7    0    5G  0 lvm  /mnt/lv_thin3
> vdc                       253:32   0    8G  0 disk
> `-vdc1                    253:33   0    8G  0 part
> vdd                       253:48   0  366K  0 disk
> vde                       253:64   0    8G  0 disk
> |-vde1                    253:65   0    8G  0 part /
> `-vde2                    253:66   0    8M  0 part
>
> Now rebooting to test the reported issue.
> But after reboot all looks sane, lsblk, lvdisplay, dev/mapper they all are
> as expected.
>
> Of course things are not mounted, but I didn't create any fstab entries, so
> for dev in lv_normal lv_thin lv_thin2 lv_thin3; do sudo mount
> /dev/mapper/vg-${dev} /mnt/${dev}; done
>
> All mount just fine and the file is still there.
>
> You reported "I was able to replicate the issue. I created VG and LV and
> rebooted the system. Found that devicemapper entries for corresponding
> devices are missing after reboot. However, LVM commands like 'vgdisplay' and
> 'lvdisplay' show proper info, but 'lsblk' doesn't show the device's LVM
> related info after reboot."
> Could you either
> 1. report your exact steps on a fresh system to cause this
> or
> 2. modify the steps I reported until the issue shows up
>
> Internal: for anyone else debugging on this, the mentioned test setup is on
> Diamond in guest "paelzer-testlvm-xenial"

Any update

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1657646

Title:
  VG/LV are not available in /dev/mapper/ after reboot

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/docker.io/+bug/1657646/+subscriptions

-- 
ubuntu-bugs mailing list
[email protected]
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to