[Kernel-packages] [Bug 1903920] Re: Warning message about cryptsetup during boot
This had been happening for me over a year ago, maybe, but seemed to have been fixed. I do have `RESUME=none` in `/etc/initramfs- tools/conf.d/resume`. Recently, this seems to have returned. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1903920 Title: Warning message about cryptsetup during boot Status in linux package in Ubuntu: Confirmed Bug description: I have noticed that since upgrade to 20.04.1 (from 18.04), a warning message appears during boot: "cryptsetup: WARNING: Option 'size' missing in crypttab for plain dm-crypt mapping cryptswap1. Please read /usr/share/doc/cryptsetup-initramfs/README.initramfs.gz and add the correct 'size' option to your crypttab(5).". AFAICT, there is no problem however. I use ecryptfs to encrypt my home directory but I don't think it's related (since the message is related to swap). This message used to appear during updates under 18.04 but I never paid attention to it. I probably use the same setup since 16.04. Others have reported the problem (https://askubuntu.com/questions/1109463, https://medium.com/@anirudha/fixing-cryptsetup-wating-for-encrypted- source-device-swapfile-error-on-linux-boot-ubuntu-20-04-5483e0dcf9e9). From what I read: - The answers indicate to edit /etc/crypttab and add a value for size but I don't know what value to use. - It seems that before 20.04, just editing /etc/crypttab was enough but some users say that they need to manually regenerate the initramfs image after each update which is boring. I have asked the question at https://answers.launchpad.net/ubuntu/+question/693615 and @bernard010 suggested to open a bug report. I have found anything worth in dmesg though. As I said everything seems fine (boot proceeds normally, I can use the computer) but I expect that there is a way to permanently make this warning disappear. ProblemType: Bug DistroRelease: Ubuntu 20.04 Package: linux-image-5.4.0-53-generic 5.4.0-53.59 ProcVersionSignature: Ubuntu 5.4.0-53.59-generic 5.4.65 Uname: Linux 5.4.0-53-generic x86_64 ApportVersion: 2.20.11-0ubuntu27.11 Architecture: amd64 AudioDevicesInUse: USERPID ACCESS COMMAND /dev/snd/controlC0: mathieu2322 F pulseaudio CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Date: Wed Nov 11 23:02:47 2020 EcryptfsInUse: Yes InstallationDate: Installed on 2017-01-06 (1405 days ago) InstallationMedia: Ubuntu 16.04.1 LTS "Xenial Xerus" - Release amd64 (20160719) MachineType: Hewlett-Packard HP EliteBook 2570p ProcFB: 0 i915drmfb ProcKernelCmdLine: BOOT_IMAGE=/boot/vmlinuz-5.4.0-53-generic root=UUID=74d77edf-b4b7-4815-865f-b7602f3396b4 ro quiet splash vt.handoff=7 RelatedPackageVersions: linux-restricted-modules-5.4.0-53-generic N/A linux-backports-modules-5.4.0-53-generic N/A linux-firmware1.187.4 SourcePackage: linux UpgradeStatus: Upgraded to focal on 2020-10-04 (38 days ago) dmi.bios.date: 02/01/2013 dmi.bios.vendor: Hewlett-Packard dmi.bios.version: 68ISB Ver. F.40 dmi.board.name: 17DF dmi.board.vendor: Hewlett-Packard dmi.board.version: KBC Version 61.21 dmi.chassis.asset.tag: CNU3229ST9 dmi.chassis.type: 10 dmi.chassis.vendor: Hewlett-Packard dmi.modalias: dmi:bvnHewlett-Packard:bvr68ISBVer.F.40:bd02/01/2013:svnHewlett-Packard:pnHPEliteBook2570p:pvrA1029C1102:rvnHewlett-Packard:rn17DF:rvrKBCVersion61.21:cvnHewlett-Packard:ct10:cvr: dmi.product.family: 103C_5336AN G=N L=BUS B=HP S=ELI dmi.product.name: HP EliteBook 2570p dmi.product.sku: C9S65EC#ABF dmi.product.version: A1029C1102 dmi.sys.vendor: Hewlett-Packard To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1903920/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
I don't believe `fimd` is sufficient, but running `stat` on a file was enough to trigger it for me. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in linux package in Ubuntu: Invalid Status in ubuntu-release-upgrader package in Ubuntu: Confirmed Status in zfs-linux package in Ubuntu: Fix Released Status in linux source package in Impish: Fix Released Status in ubuntu-release-upgrader source package in Impish: Confirmed Status in zfs-linux source package in Impish: Fix Released Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
Thanks, Simon. I'll verify a revert from the menu after I update and post there if things don't go well. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: Fix Released Status in zfs-linux source package in Impish: Fix Released Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
If the issue doesn't affect hirsute then it should be possible to update an impish system by chroot from hirsute media, I would think. I've been booting up only occasionally this past week when I need to access something, but I find that apt, or possibly initramfs, updates seem to trigger the issue for me and render my system unbootable. I've been recovering with zfs rollback from hirsute media. Unfortunately, something also seems to go wrong with the zsys recovery entries in grub, where I end up with a dataset called rpool/ROOT/ubuntu_ (no trailing unique ID). I'll file that as a separate bug if it persists after this is resolved, though. Anyway, I plan to try to chroot and update later and I'll report back. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: Fix Released Status in zfs-linux source package in Impish: Fix Released Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
Actually, I realize I may have been overconfident about that. Should it work to fix from a chroot or is the fix in the userland tools such that a chroot from hirsute would not help due to an issue in the userland ZFS tools under the chroot? -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: Fix Released Status in zfs-linux source package in Impish: Fix Released Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1921594] Re: zfsutils-linux misspells zfs-zed recommendation
My apologies, but I tried for several hours to figure out how to submit a fix and I cannot figure it out, so I'm just reporting it as a bug. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1921594 Title: zfsutils-linux misspells zfs-zed recommendation Status in zfs-linux package in Ubuntu: New Bug description: Since the version 2.0.2 imports of zfs-linux, the zfsutils-linux package contains a misspelling in its recommended dependency on zfs- zed, which is spelled instead as "fs-zed". Debian has since imported Ubuntu changes into the control.ubuntu file upstream, without the typo, but the most recent 2.0.2-1ubuntu4 still has this error. I don't know how essential zed is, but the ubiquity installer does not mark zfs-zed as manually installed, so it won't be installed for new users when I think it otherwise would be. To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1921594/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1921594] [NEW] zfsutils-linux misspells zfs-zed recommendation
Public bug reported: Since the version 2.0.2 imports of zfs-linux, the zfsutils-linux package contains a misspelling in its recommended dependency on zfs-zed, which is spelled instead as "fs-zed". Debian has since imported Ubuntu changes into the control.ubuntu file upstream, without the typo, but the most recent 2.0.2-1ubuntu4 still has this error. I don't know how essential zed is, but the ubiquity installer does not mark zfs-zed as manually installed, so it won't be installed for new users when I think it otherwise would be. ** Affects: zfs-linux (Ubuntu) Importance: Undecided Status: New -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1921594 Title: zfsutils-linux misspells zfs-zed recommendation Status in zfs-linux package in Ubuntu: New Bug description: Since the version 2.0.2 imports of zfs-linux, the zfsutils-linux package contains a misspelling in its recommended dependency on zfs- zed, which is spelled instead as "fs-zed". Debian has since imported Ubuntu changes into the control.ubuntu file upstream, without the typo, but the most recent 2.0.2-1ubuntu4 still has this error. I don't know how essential zed is, but the ubiquity installer does not mark zfs-zed as manually installed, so it won't be installed for new users when I think it otherwise would be. To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1921594/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1805982] Re: Please consider making the zed daemon activated on demand
My mistake, it seems the latest import just introduced a typo and the zfsutils-linux package now depends on the (non-existent) fs-zed package, rather than zfs-zed. I'll report that as a separate bug. This one is still valid. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1805982 Title: Please consider making the zed daemon activated on demand Status in zfs-linux package in Ubuntu: New Bug description: Hi, currently if one wants most common FS tools to be able to mount them to e.g. mount them "just in case" or when considering a more capable base image or install CD then there is the problem that installing "zfsutils-linux" pulls in "zfs-zed" as a recommends. Having it on disk would be ok, but IMHO there is no reason this daemon needs to run until the first ZFS element is created. ● zfs-zed.service - ZFS Event Daemon (zed) Loaded: loaded (/lib/systemd/system/zfs-zed.service; enabled; vendor preset: enabled) Active: active (running) since Thu 2018-11-29 16:09:04 UTC; 3min 26s ago Docs: man:zed(8) Main PID: 2968 (zed) Tasks: 2 (limit: 546) Memory: 436.0K CGroup: /system.slice/zfs-zed.service └─2968 /usr/sbin/zed -F Nov 29 16:09:04 disco-zfstest systemd[1]: Started ZFS Event Daemon (zed). Nov 29 16:09:04 disco-zfstest zed[2968]: ZFS Event Daemon 0.7.9-3ubuntu6 (PID 2968) Nov 29 16:09:04 disco-zfstest zed[2968]: Processing events since eid=0 People in general would want the number of default running services low. Further than the general "not too much default daemons" there are a few issues like [1] which makes this worse if one is affected. There is no related upstream feature yet that I'd know of, nor did I find a trivial way to e.g. match a systemd unit Condition (which also is only evaluated only once, so it would miss later zfs creation) or a socket (what would it wait for). I wanted to ask here if you'd have more/better ideas or if we should directly file this upstream and then link that bug here? [1]: https://github.com/zfsonlinux/zfs/issues/7995 To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1805982/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
I have not yet had any problems with 2.0.2. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: In Progress Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1805982] Re: Please consider making the zed daemon activated on demand
It seems like this may no longer be the case. In fact, a recent upgrade on my system removed zfs-zed, which seems like maybe a mistake since the zsys-setup script still seems to expect it. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1805982 Title: Please consider making the zed daemon activated on demand Status in zfs-linux package in Ubuntu: New Bug description: Hi, currently if one wants most common FS tools to be able to mount them to e.g. mount them "just in case" or when considering a more capable base image or install CD then there is the problem that installing "zfsutils-linux" pulls in "zfs-zed" as a recommends. Having it on disk would be ok, but IMHO there is no reason this daemon needs to run until the first ZFS element is created. ● zfs-zed.service - ZFS Event Daemon (zed) Loaded: loaded (/lib/systemd/system/zfs-zed.service; enabled; vendor preset: enabled) Active: active (running) since Thu 2018-11-29 16:09:04 UTC; 3min 26s ago Docs: man:zed(8) Main PID: 2968 (zed) Tasks: 2 (limit: 546) Memory: 436.0K CGroup: /system.slice/zfs-zed.service └─2968 /usr/sbin/zed -F Nov 29 16:09:04 disco-zfstest systemd[1]: Started ZFS Event Daemon (zed). Nov 29 16:09:04 disco-zfstest zed[2968]: ZFS Event Daemon 0.7.9-3ubuntu6 (PID 2968) Nov 29 16:09:04 disco-zfstest zed[2968]: Processing events since eid=0 People in general would want the number of default running services low. Further than the general "not too much default daemons" there are a few issues like [1] which makes this worse if one is affected. There is no related upstream feature yet that I'd know of, nor did I find a trivial way to e.g. match a systemd unit Condition (which also is only evaluated only once, so it would miss later zfs creation) or a socket (what would it wait for). I wanted to ask here if you'd have more/better ideas or if we should directly file this upstream and then link that bug here? [1]: https://github.com/zfsonlinux/zfs/issues/7995 To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1805982/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
I've install zfs-dkms to ensure that I'm using the 2.0.2 kmod. I'll post back later in the week about whether I see any panics with regular use. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: In Progress Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
I'm trying out the ZFS 2.0.2 userspace packages now, from hursute (not -proposed), and will report back. I note that the kernel module for linux-modules-5.10.0-14-lowlatency that I'm using is still 2.0.1. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: In Progress Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
2.0.2 is now in hirsute-proposed. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in zfs-linux package in Ubuntu: In Progress Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1881701] Re: NVIDIA cards should have set AllowExternalGpus in /usr/share/X11/xorg.conf.d/10-nvidia.conf
** Also affects: nvidia-graphics-drivers-450 (Ubuntu) Importance: Undecided Status: New -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to nvidia-graphics-drivers-440 in Ubuntu. https://bugs.launchpad.net/bugs/1881701 Title: NVIDIA cards should have set AllowExternalGpus in /usr/share/X11/xorg.conf.d/10-nvidia.conf Status in nvidia-graphics-drivers-440 package in Ubuntu: Confirmed Status in nvidia-graphics-drivers-450 package in Ubuntu: New Bug description: I have a Lenovo Thunderbold 3 Graphics Dock (comes integrated NVIDIA MTX 1050 graphics card), see https://support.lenovo.com/de/en/accessories/acc500012 *What happened:* With nvidia-graphics-drivers-440 I just got a black screen. After looink into the /var/log/Xorg.0.log I noticed: ``` (WW) NVIDIA(GPU-0): This device is an external GPU, but external GPUs have not (WW) NVIDIA(GPU-0): been enabled with AllowExternalGpus. Disabling this device (WW) NVIDIA(GPU-0): to prevent crashes from accidental removal. (EE) NVIDIA(0): Failing initialization of X screen 0 ``` Some googling turned up https://forums.developer.nvidia.com/t/egpu- and-intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4 After I added the `Option "AllowExternalGpus" "true"` to `Section "OutputClass"` of `/usr/share/X11/xorg.conf.d/10-nvidia.conf` and rebooting the card worked. This is a matter for all eGPUs (external GPUs) that run NVIDIA I guess. See reports at https://www.reddit.com/r/thinkpad/comments/fcuhkn/running_thunderbold_3_graphics_dock_nvidia_mtx/ and aforementioned https://forums.developer.nvidia.com/t/egpu-and- intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4 *Possible solution* Add this line to /usr/share/X11/xorg.conf.d/10-nvidia.conf: ``` Section "OutputClass" Identifier "nvidia" MatchDriver "nvidia-drm" Driver "nvidia" Option "AllowEmptyInitialConfiguration" # v ADD THIS LINE v Option "AllowExternalGpus" "true" # ^ ADD THIS LINE ^^^ ModulePath "/usr/lib/x86_64-linux-gnu/nvidia/xorg" EndSection ``` ProblemType: Bug DistroRelease: Ubuntu 20.04 Package: xserver-xorg-video-nvidia-440 440.82+really.440.64-0ubuntu6 [modified: usr/share/X11/xorg.conf.d/10-nvidia.conf] ProcVersionSignature: Ubuntu 5.4.0-33.37-generic 5.4.34 Uname: Linux 5.4.0-33-generic x86_64 NonfreeKernelModules: nvidia_modeset nvidia ApportVersion: 2.20.11-0ubuntu27.2 Architecture: amd64 CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Date: Tue Jun 2 07:30:47 2020 InstallationDate: Installed on 2019-09-19 (257 days ago) InstallationMedia: Ubuntu 19.04 "Disco Dingo" - Release amd64 (20190416) ProcEnviron: TERM=xterm-256color PATH=(custom, no user) XDG_RUNTIME_DIR= LANG=de_DE.UTF-8 SHELL=/bin/bash SourcePackage: nvidia-graphics-drivers-440 UpgradeStatus: Upgraded to focal on 2020-05-29 (3 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-drivers-440/+bug/1881701/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1872108] Re: Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal
Benjamin Poirier seems to be working to upstream changes that would fix this issue. See these links for details: https://github.com/gobenji/thinkpad-x1-gen7-sound https://lore.kernel.org/alsa-devel/20200703080005.8942-1-benjamin.poir...@gmail.com/ https://lore.kernel.org/alsa-devel/20200703072302.16876-1-benjamin.poir...@gmail.com/ -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to alsa-ucm-conf in Ubuntu. https://bugs.launchpad.net/bugs/1872108 Title: Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal Status in alsa-ucm-conf package in Ubuntu: Confirmed Status in linux package in Ubuntu: Confirmed Bug description: Related to https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1845677/comments/14 Audio works and the volume seems reasonable however the audio is noticeably missing bass giving everything a high treble sound. ProblemType: Bug DistroRelease: Ubuntu 20.04 Package: alsa-ucm-conf 1.2.2-1 ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27 Uname: Linux 5.4.0-21-generic x86_64 ApportVersion: 2.20.11-0ubuntu26 Architecture: amd64 CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Date: Fri Apr 10 09:17:33 2020 Dependencies: InstallationDate: Installed on 2020-04-08 (1 days ago) InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402) PackageArchitecture: all ProcEnviron: TERM=xterm-256color PATH=(custom, no user) XDG_RUNTIME_DIR= LANG=en_US.UTF-8 SHELL=/bin/bash SourcePackage: alsa-ucm-conf UpgradeStatus: No upgrade log present (probably fresh install) --- ProblemType: Bug ApportVersion: 2.20.11-0ubuntu26 Architecture: amd64 CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Dependencies: DistroRelease: Ubuntu 20.04 InstallationDate: Installed on 2020-04-08 (1 days ago) InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402) Package: linux PackageArchitecture: all ProcEnviron: TERM=xterm-256color PATH=(custom, no user) XDG_RUNTIME_DIR= LANG=en_US.UTF-8 SHELL=/bin/bash ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27 Tags: focal Uname: Linux 5.4.0-21-generic x86_64 UpgradeStatus: No upgrade log present (probably fresh install) UserGroups: adm cdrom dip lpadmin lxd plugdev sambashare sudo _MarkForUpload: True To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/alsa-ucm-conf/+bug/1872108/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1872108] Re: Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal
This makes a world of difference. It seems to make the Headphone volume control the bottom speakers while the Speaker volume still controls the top speaker. Does anyone know enough about ALSA to know if there's a way the command could be tweaked so that the Speaker volume controls both? -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to alsa-ucm-conf in Ubuntu. https://bugs.launchpad.net/bugs/1872108 Title: Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal Status in alsa-ucm-conf package in Ubuntu: Confirmed Status in linux package in Ubuntu: Confirmed Bug description: Related to https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1845677/comments/14 Audio works and the volume seems reasonable however the audio is noticeably missing bass giving everything a high treble sound. ProblemType: Bug DistroRelease: Ubuntu 20.04 Package: alsa-ucm-conf 1.2.2-1 ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27 Uname: Linux 5.4.0-21-generic x86_64 ApportVersion: 2.20.11-0ubuntu26 Architecture: amd64 CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Date: Fri Apr 10 09:17:33 2020 Dependencies: InstallationDate: Installed on 2020-04-08 (1 days ago) InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402) PackageArchitecture: all ProcEnviron: TERM=xterm-256color PATH=(custom, no user) XDG_RUNTIME_DIR= LANG=en_US.UTF-8 SHELL=/bin/bash SourcePackage: alsa-ucm-conf UpgradeStatus: No upgrade log present (probably fresh install) --- ProblemType: Bug ApportVersion: 2.20.11-0ubuntu26 Architecture: amd64 CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Dependencies: DistroRelease: Ubuntu 20.04 InstallationDate: Installed on 2020-04-08 (1 days ago) InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402) Package: linux PackageArchitecture: all ProcEnviron: TERM=xterm-256color PATH=(custom, no user) XDG_RUNTIME_DIR= LANG=en_US.UTF-8 SHELL=/bin/bash ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27 Tags: focal Uname: Linux 5.4.0-21-generic x86_64 UpgradeStatus: No upgrade log present (probably fresh install) UserGroups: adm cdrom dip lpadmin lxd plugdev sambashare sudo _MarkForUpload: True To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/alsa-ucm-conf/+bug/1872108/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1790515] Re: black screen on ubuntu GTX 1050 eGPU
Could you try the fix from #1881701, please? https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics- drivers-440/+bug/1881701 -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to nvidia-graphics-drivers-390 in Ubuntu. https://bugs.launchpad.net/bugs/1790515 Title: black screen on ubuntu GTX 1050 eGPU Status in nvidia-graphics-drivers-390 package in Ubuntu: New Bug description: I'm using a laptop lenovo thinkpad T410 with intel graphics and egpu + GTX 1050. (using expresscard or mini pcie both failed) When I install the 390 driver my graphical display fails. I know that egpu is not the best solution but for some people it works. I tried: 390, 396, 384 nvidia drivers from ppa and from NVIDIA (.run) Ubuntu and Xubuntu (18, 17, 16) using gdm3 and lightgdm blacklist nouveau and everything similar Sometimes Ubuntu 18.04 freeze using live usb. To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-drivers-390/+bug/1790515/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1881701] Re: NVIDIA cards should have set AllowExternalGpus in /usr/share/X11/xorg.conf.d/10-nvidia.conf
I can confirm that adding: Option "AllowExternalGpus" Fixes my issue. This may be the same as #1790515, too. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to nvidia-graphics-drivers-440 in Ubuntu. https://bugs.launchpad.net/bugs/1881701 Title: NVIDIA cards should have set AllowExternalGpus in /usr/share/X11/xorg.conf.d/10-nvidia.conf Status in nvidia-graphics-drivers-440 package in Ubuntu: New Bug description: I have a Lenovo Thunderbold 3 Graphics Dock (comes integrated NVIDIA MTX 1050 graphics card), see https://support.lenovo.com/de/en/accessories/acc500012 *What happened:* With nvidia-graphics-drivers-440 I just got a black screen. After looink into the /var/log/Xorg.0.log I noticed: ``` (WW) NVIDIA(GPU-0): This device is an external GPU, but external GPUs have not (WW) NVIDIA(GPU-0): been enabled with AllowExternalGpus. Disabling this device (WW) NVIDIA(GPU-0): to prevent crashes from accidental removal. (EE) NVIDIA(0): Failing initialization of X screen 0 ``` Some googling turned up https://forums.developer.nvidia.com/t/egpu- and-intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4 After I added the `Option "AllowExternalGpus" "true"` to `Section "OutputClass"` of `/usr/share/X11/xorg.conf.d/10-nvidia.conf` and rebooting the card worked. This is a matter for all eGPUs (external GPUs) that run NVIDIA I guess. See reports at https://www.reddit.com/r/thinkpad/comments/fcuhkn/running_thunderbold_3_graphics_dock_nvidia_mtx/ and aforementioned https://forums.developer.nvidia.com/t/egpu-and- intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4 *Possible solution* Add this line to /usr/share/X11/xorg.conf.d/10-nvidia.conf: ``` Section "OutputClass" Identifier "nvidia" MatchDriver "nvidia-drm" Driver "nvidia" Option "AllowEmptyInitialConfiguration" # v ADD THIS LINE v Option "AllowExternalGpus" "true" # ^ ADD THIS LINE ^^^ ModulePath "/usr/lib/x86_64-linux-gnu/nvidia/xorg" EndSection ``` ProblemType: Bug DistroRelease: Ubuntu 20.04 Package: xserver-xorg-video-nvidia-440 440.82+really.440.64-0ubuntu6 [modified: usr/share/X11/xorg.conf.d/10-nvidia.conf] ProcVersionSignature: Ubuntu 5.4.0-33.37-generic 5.4.34 Uname: Linux 5.4.0-33-generic x86_64 NonfreeKernelModules: nvidia_modeset nvidia ApportVersion: 2.20.11-0ubuntu27.2 Architecture: amd64 CasperMD5CheckResult: skip CurrentDesktop: ubuntu:GNOME Date: Tue Jun 2 07:30:47 2020 InstallationDate: Installed on 2019-09-19 (257 days ago) InstallationMedia: Ubuntu 19.04 "Disco Dingo" - Release amd64 (20190416) ProcEnviron: TERM=xterm-256color PATH=(custom, no user) XDG_RUNTIME_DIR= LANG=de_DE.UTF-8 SHELL=/bin/bash SourcePackage: nvidia-graphics-drivers-440 UpgradeStatus: Upgraded to focal on 2020-05-29 (3 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-drivers-440/+bug/1881701/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root
Confirmed fixed over here. Thanks for the fast turnaround! -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1673197 Title: Upgrading zfs-initramfs breaks booting from a zfs root Status in zfs-linux package in Ubuntu: Fix Released Status in zfs-linux package in Debian: Fix Released Bug description: Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot. I've EFI and a small ext2 boot partition, and the rest of my SDD is a zfs pool. After upgrade I'm dropped in the initramfs shell and no zpools are available. Downgrading to 0.6.5.8-0ubuntu9 makes it boot again. ProblemType: Bug DistroRelease: Ubuntu 17.04 Package: zfs-initramfs 0.6.5.9-4ubuntu1 ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1 Uname: Linux 4.10.0-11-generic x86_64 NonfreeKernelModules: zfs zunicode zavl zcommon znvpair ApportVersion: 2.20.4-0ubuntu2 Architecture: amd64 CurrentDesktop: Unity:Unity7 Date: Wed Mar 15 18:29:04 2017 InstallationDate: Installed on 2016-05-06 (312 days ago) InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 (20160420.1) PackageArchitecture: all SourcePackage: zfs-linux UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root
My last comment got cut off. I was going to say I don't see /usr/share /initramfs-tools/conf.d/zfs on my filesystem. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1673197 Title: Upgrading zfs-initramfs breaks booting from a zfs root Status in zfs-linux package in Ubuntu: In Progress Bug description: Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot. I've EFI and a small ext2 boot partition, and the rest of my SDD is a zfs pool. After upgrade I'm dropped in the initramfs shell and no zpools are available. Downgrading to 0.6.5.8-0ubuntu9 makes it boot again. ProblemType: Bug DistroRelease: Ubuntu 17.04 Package: zfs-initramfs 0.6.5.9-4ubuntu1 ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1 Uname: Linux 4.10.0-11-generic x86_64 NonfreeKernelModules: zfs zunicode zavl zcommon znvpair ApportVersion: 2.20.4-0ubuntu2 Architecture: amd64 CurrentDesktop: Unity:Unity7 Date: Wed Mar 15 18:29:04 2017 InstallationDate: Installed on 2016-05-06 (312 days ago) InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 (20160420.1) PackageArchitecture: all SourcePackage: zfs-linux UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root
Maybe this is some regression in zfs-initramfs? I see some mention of this in https://bugs.launchpad.net/ubuntu/+source/zfs-initramfs/+bug/1530953 I don -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1673197 Title: Upgrading zfs-initramfs breaks booting from a zfs root Status in zfs-linux package in Ubuntu: In Progress Bug description: Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot. I've EFI and a small ext2 boot partition, and the rest of my SDD is a zfs pool. After upgrade I'm dropped in the initramfs shell and no zpools are available. Downgrading to 0.6.5.8-0ubuntu9 makes it boot again. ProblemType: Bug DistroRelease: Ubuntu 17.04 Package: zfs-initramfs 0.6.5.9-4ubuntu1 ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1 Uname: Linux 4.10.0-11-generic x86_64 NonfreeKernelModules: zfs zunicode zavl zcommon znvpair ApportVersion: 2.20.4-0ubuntu2 Architecture: amd64 CurrentDesktop: Unity:Unity7 Date: Wed Mar 15 18:29:04 2017 InstallationDate: Installed on 2016-05-06 (312 days ago) InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 (20160420.1) PackageArchitecture: all SourcePackage: zfs-linux UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root
I just happened to find that parameter mentioned in some ZFS root setup guides and tried it as a wild guess. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1673197 Title: Upgrading zfs-initramfs breaks booting from a zfs root Status in zfs-linux package in Ubuntu: In Progress Bug description: Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot. I've EFI and a small ext2 boot partition, and the rest of my SDD is a zfs pool. After upgrade I'm dropped in the initramfs shell and no zpools are available. Downgrading to 0.6.5.8-0ubuntu9 makes it boot again. ProblemType: Bug DistroRelease: Ubuntu 17.04 Package: zfs-initramfs 0.6.5.9-4ubuntu1 ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1 Uname: Linux 4.10.0-11-generic x86_64 NonfreeKernelModules: zfs zunicode zavl zcommon znvpair ApportVersion: 2.20.4-0ubuntu2 Architecture: amd64 CurrentDesktop: Unity:Unity7 Date: Wed Mar 15 18:29:04 2017 InstallationDate: Installed on 2016-05-06 (312 days ago) InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 (20160420.1) PackageArchitecture: all SourcePackage: zfs-linux UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root
I don't know what changed, but adding `boot=zfs` to my grub kernel parameters fixes this. I never needed it in the past and `update-grub` doesn't create it. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1673197 Title: Upgrading zfs-initramfs breaks booting from a zfs root Status in zfs-linux package in Ubuntu: In Progress Bug description: Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot. I've EFI and a small ext2 boot partition, and the rest of my SDD is a zfs pool. After upgrade I'm dropped in the initramfs shell and no zpools are available. Downgrading to 0.6.5.8-0ubuntu9 makes it boot again. ProblemType: Bug DistroRelease: Ubuntu 17.04 Package: zfs-initramfs 0.6.5.9-4ubuntu1 ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1 Uname: Linux 4.10.0-11-generic x86_64 NonfreeKernelModules: zfs zunicode zavl zcommon znvpair ApportVersion: 2.20.4-0ubuntu2 Architecture: amd64 CurrentDesktop: Unity:Unity7 Date: Wed Mar 15 18:29:04 2017 InstallationDate: Installed on 2016-05-06 (312 days ago) InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 (20160420.1) PackageArchitecture: all SourcePackage: zfs-linux UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago) To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp
[Kernel-packages] [Bug 1289797] Re: [LENOVO 20A8A01DLM] suspend/resume failure
Confirm fixed by the most recent BIOS update. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1289797 Title: [LENOVO 20A8A01DLM] suspend/resume failure Status in “linux” package in Ubuntu: Triaged Bug description: It dies when I suspend it, the adaptive key strip and the power key keep turned on. ProblemType: KernelOops DistroRelease: Ubuntu 14.04 Package: linux-image-3.13.0-16-generic 3.13.0-16.36 ProcVersionSignature: Ubuntu 3.13.0-16.36-generic 3.13.5 Uname: Linux 3.13.0-16-generic i686 Annotation: This occured during a previous suspend and prevented it from resuming properly. ApportVersion: 2.13.2-0ubuntu5 Architecture: i386 AudioDevicesInUse: USERPID ACCESS COMMAND /dev/snd/controlC0: ubuntu 3349 F pulseaudio /dev/snd/controlC1: ubuntu 3349 F pulseaudio CRDA: country MX: (2402 - 2472 @ 40), (3, 27) (5170 - 5250 @ 40), (3, 17) (5250 - 5330 @ 40), (3, 23), DFS (5735 - 5835 @ 40), (3, 30) CasperVersion: 1.337 Date: Fri Mar 7 21:10:38 2014 ExecutablePath: /usr/share/apport/apportcheckresume Failure: suspend/resume InterpreterPath: /usr/bin/python3.4 LiveMediaBuild: Ubuntu 14.04 LTS "Trusty Tahr" - Alpha i386 (20140307) MachineType: LENOVO 20A8A01DLM ProcCmdline: /usr/bin/python3 /usr/share/apport/apportcheckresume ProcEnviron: TERM=linux PATH=(custom, no user) ProcFB: 0 inteldrmfb ProcKernelCmdLine: noprompt cdrom-detect/try-usb=true persistent file=/cdrom/preseed/hostname.seed boot=casper initrd=/casper/initrd.lz quiet splash -- maybe-ubiquity PulseList: Error: command ['pacmd', 'list'] failed with exit code 1: No PulseAudio daemon running, or not running as session daemon. RelatedPackageVersions: linux-restricted-modules-3.13.0-16-generic N/A linux-backports-modules-3.13.0-16-generic N/A linux-firmware 1.126 SourcePackage: linux Title: [LENOVO 20A8A01DLM] suspend/resume failure UpgradeStatus: No upgrade log present (probably fresh install) UserGroups: dmi.bios.date: 12/03/2013 dmi.bios.vendor: LENOVO dmi.bios.version: GRET29WW (1.06 ) dmi.board.asset.tag: Not Available dmi.board.name: 20A8A01DLM dmi.board.vendor: LENOVO dmi.board.version: SDK0E50510 Pro dmi.chassis.asset.tag: No Asset Information dmi.chassis.type: 10 dmi.chassis.vendor: LENOVO dmi.chassis.version: Not Available dmi.modalias: dmi:bvnLENOVO:bvrGRET29WW(1.06):bd12/03/2013:svnLENOVO:pn20A8A01DLM:pvrThinkPadX1Carbon2nd:rvnLENOVO:rn20A8A01DLM:rvrSDK0E50510Pro:cvnLENOVO:ct10:cvrNotAvailable: dmi.product.name: 20A8A01DLM dmi.product.version: ThinkPad X1 Carbon 2nd dmi.sys.vendor: LENOVO To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1289797/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp