It is my own desktop, so I can do some experiments. I have tried numerous versions and have found that the last version it works on (with no modifications on a fresh install) is 21.10, I tried 20.04 (with zero updates) and 20.04.4 (with and without updates) and both of them exhibit the same result as 24.04. So the break is somewhere between Impish and Jammy.
I appreciate all the help :) Let me know if there is anything else I can be testing. ** Description changed: - On an almost fresh install of Ubuntu 24.04, with dmraid installed. When booting the RAID array is detected and up, but the partitions inside of it are not detected and I have to manually run `sudo kpartx -a /dev/mapper/isw_caiggcbbj_Volume1` to make them appear correctly. - This is a regression from Ubuntu 20.04, where it just works at boot. + On an almost fresh install of Ubuntu any version between 22.04 to 24.04, with dmraid installed. When booting the RAID array is detected and up, but the partitions inside of it are not detected and I have to manually run `sudo kpartx -a /dev/mapper/isw_caiggcbbj_Volume1` to make them appear correctly. + This is a regression, as from Ubuntu 20.04 to 21.10, it just works at boot. Seemingly kpartx-boot is what is meant to make it just work at boot. --- MORE INFO --- It is the same PC, I have the installs side by side on the same boot disk (to clarify, my boot drive is not under RAID). I am using Intel Rapid Storage Technology configured in my BIOS and dmraid (I briefly tried mdadm, but it didn't work in my 15 minutes of trying). A summary of the difference between the two OSes is at the bottom. ## For Ubuntu 20.04, where it is working ## `dmsetup --version`: ``` Library version: 1.02.167 (2019-11-30) Driver version: 4.45.0 ``` `dmraid --version`: ``` dmraid version: 1.0.0.rc16 (2009.09.16) shared dmraid library version: 1.0.0.rc16 (2009.09.16) device-mapper version: 4.45.0 ``` `lsmod | grep 'raid'` produces no output. `lsmod | grep 'dm_'`: ``` dm_mirror 24576 0 dm_region_hash 24576 1 dm_mirror dm_log 20480 2 dm_region_hash,dm_mirror ``` `dmraid -r`: ``` /dev/sda: isw, "isw_caiggcbbj", GROUP, ok, 468862126 sectors, data@ 0 /dev/sdb: isw, "isw_caiggcbbj", GROUP, ok, 468862126 sectors, data@ 0 ``` `dmraid -s`: ``` *** Group superset isw_caiggcbbj --> Active Subset name : isw_caiggcbbj_Volume1 size : 937714176 stride : 128 type : stripe status : ok subsets: 0 devs : 2 spares : 0 ``` `dmsetup ls`: ``` isw_caiggcbbj_Volume1 (253:0) isw_caiggcbbj_Volume1p1 (253:2) ``` `ls -l /dev/mapper/`: ``` total 0 crw------- 1 root root 10, 236 Jul 31 19:40 control brw-rw---- 1 root disk 253, 1 Jul 31 19:40 isw_caiggcbbj_Volume1 lrwxrwxrwx 1 root root 7 Jul 31 19:40 isw_caiggcbbj_Volume1p1 -> ../dm-2 ``` `lsblk`: ``` NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 223.6G 0 disk └─isw_caiggcbbj_Volume1 253:1 0 447.1G 0 dmraid └─isw_caiggcbbj_Volume1p1 253:2 0 447.1G 0 part /media/ben/SSD_ext4 sdb 8:16 0 223.6G 0 disk └─isw_caiggcbbj_Volume1 253:1 0 447.1G 0 dmraid └─isw_caiggcbbj_Volume1p1 253:2 0 447.1G 0 part /media/ben/SSD_ext4 ``` `fdisk -l`: ``` Disk /dev/sda: 223.58 GiB, 240057409536 bytes, 468862128 sectors Disk model: TCSUNBOW X3 240G Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: dos Disk identifier: 0x00000000 Device Boot Start End Sectors Size Id Type /dev/sda1 1 468862127 468862127 223.6G ee GPT Disk /dev/sdb: 223.58 GiB, 240057409536 bytes, 468862128 sectors Disk model: TCSUNBOW X3 240G Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk /dev/mapper/isw_caiggcbbj_Volume1: 447.14 GiB, 480109658112 bytes, 937714176 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 65536 bytes / 131072 bytes Disklabel type: gpt Disk identifier: BEA6AF5B-DF82-4D02-BD4A-1027B795151F Device Start End Sectors Size Type /dev/mapper/isw_caiggcbbj_Volume1p1 2048 937713663 937711616 447.1G Linux filesystem ``` `journalctl -b | grep -i "raid\|dm-\d\|isw"`: ``` Jul 31 19:40:34 Ubentu-Desktop kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: [email protected] Jul 31 19:40:34 Ubentu-Desktop kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 6 ports 6 Gbps 0x3f impl RAID mode Jul 31 19:40:34 Ubentu-Desktop dmraid-activate[806]: ERROR: Cannot retrieve RAID set information for isw_caiggcbbj_Volume1 Jul 31 19:40:34 Ubentu-Desktop dmraid-activate[962]: ERROR: Cannot retrieve RAID set information for isw_caiggcbbj_Volume1 Jul 31 19:40:36 Ubentu-Desktop udisksd[1270]: failed to load module mdraid: libbd_mdraid.so.2: cannot open shared object file: No such file or directory Jul 31 19:40:36 Ubentu-Desktop udisksd[1270]: Failed to load the 'mdraid' libblockdev plugin ``` ## And then for Ubuntu 24.04, where it is not working ## `dmsetup --version`: ``` Library version: 1.02.185 (2022-05-18) Driver version: 4.48.0 ``` `dmraid --version`: ``` dmraid version: 1.0.0.rc16 (2009.09.16) shared dmraid library version: 1.0.0.rc16 (2009.09.16) device-mapper version: 4.48.0 ``` `lsmod | grep 'raid'` is the same as for Ubuntu 20.04 (no output). `lsmod | grep 'dm_'` is the same as for Ubuntu 20.04. `dmraid -r` is the same as for Ubuntu 20.04. `dmraid -s` is the same as for Ubuntu 20.04. `dmsetup ls`: ``` isw_caiggcbbj_Volume1 (252:0) ``` `ls -l /dev/mapper/`: ``` crw------- 1 root root 10, 236 Jul 31 20:11 control brw-rw---- 1 root disk 252, 0 Jul 31 20:11 isw_caiggcbbj_Volume1 ``` `lsblk`: ``` NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS sda 8:0 0 223.6G 0 disk └─isw_caiggcbbj_Volume1 252:0 0 447.1G 0 dmraid sdb 8:16 0 223.6G 0 disk └─isw_caiggcbbj_Volume1 252:0 0 447.1G 0 dmraid ``` `fdisk -l`: ``` Disk /dev/sda: 223.57 GiB, 240057409536 bytes, 468862128 sectors Disk model: TCSUNBOW X3 240G Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disklabel type: dos Disk identifier: 0x00000000 Device Boot Start End Sectors Size Id Type /dev/sda1 1 468862127 468862127 223.6G ee GPT Disk /dev/sdb: 223.57 GiB, 240057409536 bytes, 468862128 sectors Disk model: TCSUNBOW X3 240G Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk /dev/mapper/isw_caiggcbbj_Volume1: 447.14 GiB, 480109658112 bytes, 937714176 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 65536 bytes / 131072 bytes Disklabel type: gpt Disk identifier: BEA6AF5B-DF82-4D02-BD4A-1027B795151F Device Start End Sectors Size Type /dev/mapper/isw_caiggcbbj_Volume1-part1 2048 937713663 937711616 447.1G Linux filesystem ``` `journalctl -b | grep -i "raid\|dm-\d\|isw"`: ``` Jul 31 19:12:19 Ubentu-Desktop-24-04 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 6 ports 6 Gbps 0x3f impl RAID mode Jul 31 19:12:19 Ubentu-Desktop-24-04 (udev-worker)[468]: dm-0: Process '/usr/bin/unshare -m /usr/bin/snap auto-import --mount=/dev/dm-0' failed with exit code 1. Jul 31 19:12:19 Ubentu-Desktop-24-04 dmraid-activate[668]: ERROR: Cannot retrieve RAID set information for isw_caiggcbbj_Volume1 Jul 31 19:12:19 Ubentu-Desktop-24-04 dmraid-activate[711]: ERROR: Cannot retrieve RAID set information for isw_caiggcbbj_Volume1 ``` ## Summary of differences ## - Version update of device-mapper. - The partition is not showing up in 24.04 (i.e. this bug). - `fdisk -l` is reporting that the partition should be "/dev/mapper/isw_caiggcbbj_Volume1-part1" instead of "/dev/mapper/isw_caiggcbbj_Volume1p1" (suffix change) on 24.04 vs 20.04 (respectivly). After running the `kpartx` command on 24.04 this does "correct itself" and the partition is mapped as "isw_caiggcbbj_Volume1p1". - Both "/dev/dm-1" for the RAID array, but when I run the `kpartx` command on 24.04 the device for the partition becomes "dm-0" (instead of "dm-2" like on 20.04). - The major device number changes from 253 on 20.04 to 252 on 24.04. Although if I am not mistaken, this just means it's in the dynamic range, so doesn't mean anything. - There is an additional error message on 24.04 - "(udev-worker)[468]: dm-0: Process '/usr/bin/unshare -m /usr/bin/snap auto-import --mount=/dev/dm-0' failed with exit code 1". ## Things I've tried ## 1. `apt update && apt upgrade` 2. Rebooting many times (to check for an obvious race condition). 3. `update-initramfs -u` (although this was already done when installing dmraid / kpartx-boot). --- --- --- ProblemType: Bug DistroRelease: Ubuntu 24.04 Package: kpartx 0.9.4-5ubuntu8 ProcVersionSignature: Ubuntu 6.8.0-39.39-generic 6.8.8 Uname: Linux 6.8.0-39-generic x86_64 ApportVersion: 2.28.1-0ubuntu3 Architecture: amd64 CasperMD5CheckResult: pass CurrentDesktop: ubuntu:GNOME Date: Wed Jul 31 23:40:55 2024 InstallationDate: Installed on 2024-07-31 (0 days ago) InstallationMedia: Ubuntu 24.04 LTS "Noble Numbat" - Release amd64 (20240424) ProcEnviron: LANG=en_US.UTF-8 PATH=(custom, no user) SHELL=/bin/bash TERM=xterm-256color SourcePackage: multipath-tools UpgradeStatus: No upgrade log present (probably fresh install) -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/2075442 Title: RAID partitions not auto detected at boot To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/multipath-tools/+bug/2075442/+subscriptions -- ubuntu-bugs mailing list [email protected] https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs
