[Kernel-packages] [Bug 1903920] Re: Warning message about cryptsetup during boot

2022-12-25 Thread Randall Leeds
This had been happening for me over a year ago, maybe, but seemed to
have been fixed. I do have `RESUME=none` in `/etc/initramfs-
tools/conf.d/resume`.

Recently, this seems to have returned.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1903920

Title:
  Warning message about cryptsetup during boot

Status in linux package in Ubuntu:
  Confirmed

Bug description:
  I have noticed that since upgrade to 20.04.1 (from 18.04), a warning
  message appears during boot: "cryptsetup: WARNING: Option 'size'
  missing in crypttab for plain dm-crypt mapping cryptswap1. Please read
  /usr/share/doc/cryptsetup-initramfs/README.initramfs.gz and add the
  correct 'size' option to your crypttab(5).".

  AFAICT, there is no problem however. I use ecryptfs to encrypt my home
  directory but I don't think it's related (since the message is related
  to swap). This message used to appear during updates under 18.04 but I
  never paid attention to it. I probably use the same setup since 16.04.

  Others have reported the problem
  (https://askubuntu.com/questions/1109463,
  https://medium.com/@anirudha/fixing-cryptsetup-wating-for-encrypted-
  source-device-swapfile-error-on-linux-boot-ubuntu-20-04-5483e0dcf9e9).
  From what I read:

    - The answers indicate to edit /etc/crypttab and add a value for size but I 
don't know what value to use.
    - It seems that before 20.04, just editing /etc/crypttab was enough but 
some users say that they need to manually regenerate the initramfs image after 
each update which is boring.

  I have asked the question at
  https://answers.launchpad.net/ubuntu/+question/693615 and @bernard010
  suggested to open a bug report.

  I have found anything worth in dmesg though.

  As I said everything seems fine (boot proceeds normally, I can use the
  computer) but I expect that there is a way to permanently make this
  warning disappear.

  ProblemType: Bug
  DistroRelease: Ubuntu 20.04
  Package: linux-image-5.4.0-53-generic 5.4.0-53.59
  ProcVersionSignature: Ubuntu 5.4.0-53.59-generic 5.4.65
  Uname: Linux 5.4.0-53-generic x86_64
  ApportVersion: 2.20.11-0ubuntu27.11
  Architecture: amd64
  AudioDevicesInUse:
   USERPID ACCESS COMMAND
   /dev/snd/controlC0:  mathieu2322 F pulseaudio
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Date: Wed Nov 11 23:02:47 2020
  EcryptfsInUse: Yes
  InstallationDate: Installed on 2017-01-06 (1405 days ago)
  InstallationMedia: Ubuntu 16.04.1 LTS "Xenial Xerus" - Release amd64 
(20160719)
  MachineType: Hewlett-Packard HP EliteBook 2570p
  ProcFB: 0 i915drmfb
  ProcKernelCmdLine: BOOT_IMAGE=/boot/vmlinuz-5.4.0-53-generic 
root=UUID=74d77edf-b4b7-4815-865f-b7602f3396b4 ro quiet splash vt.handoff=7
  RelatedPackageVersions:
   linux-restricted-modules-5.4.0-53-generic N/A
   linux-backports-modules-5.4.0-53-generic  N/A
   linux-firmware1.187.4
  SourcePackage: linux
  UpgradeStatus: Upgraded to focal on 2020-10-04 (38 days ago)
  dmi.bios.date: 02/01/2013
  dmi.bios.vendor: Hewlett-Packard
  dmi.bios.version: 68ISB Ver. F.40
  dmi.board.name: 17DF
  dmi.board.vendor: Hewlett-Packard
  dmi.board.version: KBC Version 61.21
  dmi.chassis.asset.tag: CNU3229ST9
  dmi.chassis.type: 10
  dmi.chassis.vendor: Hewlett-Packard
  dmi.modalias: 
dmi:bvnHewlett-Packard:bvr68ISBVer.F.40:bd02/01/2013:svnHewlett-Packard:pnHPEliteBook2570p:pvrA1029C1102:rvnHewlett-Packard:rn17DF:rvrKBCVersion61.21:cvnHewlett-Packard:ct10:cvr:
  dmi.product.family: 103C_5336AN G=N L=BUS B=HP S=ELI
  dmi.product.name: HP EliteBook 2570p
  dmi.product.sku: C9S65EC#ABF
  dmi.product.version: A1029C1102
  dmi.sys.vendor: Hewlett-Packard

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1903920/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-10-19 Thread Randall Leeds
I don't believe `fimd` is sufficient, but running `stat` on a file was
enough to trigger it for me.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in linux package in Ubuntu:
  Invalid
Status in ubuntu-release-upgrader package in Ubuntu:
  Confirmed
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Impish:
  Fix Released
Status in ubuntu-release-upgrader source package in Impish:
  Confirmed
Status in zfs-linux source package in Impish:
  Fix Released

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-09-27 Thread Randall Leeds
Thanks, Simon. I'll verify a revert from the menu after I update and
post there if things don't go well.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in zfs-linux source package in Impish:
  Fix Released

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-09-27 Thread Randall Leeds
If the issue doesn't affect hirsute then it should be possible to update
an impish system by chroot from hirsute media, I would think.

I've been booting up only occasionally this past week when I need to
access something, but I find that apt, or possibly initramfs, updates
seem to trigger the issue for me and render my system unbootable. I've
been recovering with zfs rollback from hirsute media. Unfortunately,
something also seems to go wrong with the zsys recovery entries in grub,
where I end up with a dataset called rpool/ROOT/ubuntu_ (no trailing
unique ID). I'll file that as a separate bug if it persists after this
is resolved, though.

Anyway, I plan to try to chroot and update later and I'll report back.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in zfs-linux source package in Impish:
  Fix Released

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-09-27 Thread Randall Leeds
Actually, I realize I may have been overconfident about that. Should it
work to fix from a chroot or is the fix in the userland tools such that
a chroot from hirsute would not help due to an issue in the userland ZFS
tools under the chroot?

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in zfs-linux source package in Impish:
  Fix Released

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1921594] Re: zfsutils-linux misspells zfs-zed recommendation

2021-03-27 Thread Randall Leeds
My apologies, but I tried for several hours to figure out how to submit
a fix and I cannot figure it out, so I'm just reporting it as a bug.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1921594

Title:
  zfsutils-linux misspells zfs-zed recommendation

Status in zfs-linux package in Ubuntu:
  New

Bug description:
  Since the version 2.0.2 imports of zfs-linux, the zfsutils-linux
  package contains a misspelling in its recommended dependency on zfs-
  zed, which is spelled instead as "fs-zed".

  Debian has since imported Ubuntu changes into the control.ubuntu file
  upstream, without the typo, but the most recent 2.0.2-1ubuntu4 still
  has this error.

  I don't know how essential zed is, but the ubiquity installer does not
  mark zfs-zed as manually installed, so it won't be installed for new
  users when I think it otherwise would be.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1921594/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1921594] [NEW] zfsutils-linux misspells zfs-zed recommendation

2021-03-27 Thread Randall Leeds
Public bug reported:

Since the version 2.0.2 imports of zfs-linux, the zfsutils-linux package
contains a misspelling in its recommended dependency on zfs-zed, which
is spelled instead as "fs-zed".

Debian has since imported Ubuntu changes into the control.ubuntu file
upstream, without the typo, but the most recent 2.0.2-1ubuntu4 still has
this error.

I don't know how essential zed is, but the ubiquity installer does not
mark zfs-zed as manually installed, so it won't be installed for new
users when I think it otherwise would be.

** Affects: zfs-linux (Ubuntu)
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1921594

Title:
  zfsutils-linux misspells zfs-zed recommendation

Status in zfs-linux package in Ubuntu:
  New

Bug description:
  Since the version 2.0.2 imports of zfs-linux, the zfsutils-linux
  package contains a misspelling in its recommended dependency on zfs-
  zed, which is spelled instead as "fs-zed".

  Debian has since imported Ubuntu changes into the control.ubuntu file
  upstream, without the typo, but the most recent 2.0.2-1ubuntu4 still
  has this error.

  I don't know how essential zed is, but the ubiquity installer does not
  mark zfs-zed as manually installed, so it won't be installed for new
  users when I think it otherwise would be.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1921594/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1805982] Re: Please consider making the zed daemon activated on demand

2021-02-27 Thread Randall Leeds
My mistake, it seems the latest import just introduced a typo and the
zfsutils-linux package now depends on the (non-existent) fs-zed package,
rather than zfs-zed. I'll report that as a separate bug. This one is
still valid.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1805982

Title:
  Please consider making the zed daemon activated on demand

Status in zfs-linux package in Ubuntu:
  New

Bug description:
  Hi,
  currently if one wants most common FS tools to be able to mount them to e.g. 
mount them "just in case" or when considering a more capable base image or 
install CD then there is the problem that installing "zfsutils-linux" pulls in 
"zfs-zed" as a recommends.

  Having it on disk would be ok, but IMHO there is no reason this daemon
  needs to run until the first ZFS element is created.

  ● zfs-zed.service - ZFS Event Daemon (zed)
 Loaded: loaded (/lib/systemd/system/zfs-zed.service; enabled; vendor 
preset: enabled)
 Active: active (running) since Thu 2018-11-29 16:09:04 UTC; 3min 26s ago
   Docs: man:zed(8)
   Main PID: 2968 (zed)
  Tasks: 2 (limit: 546)
 Memory: 436.0K
 CGroup: /system.slice/zfs-zed.service
 └─2968 /usr/sbin/zed -F

  Nov 29 16:09:04 disco-zfstest systemd[1]: Started ZFS Event Daemon (zed).
  Nov 29 16:09:04 disco-zfstest zed[2968]: ZFS Event Daemon 0.7.9-3ubuntu6 (PID 
2968)
  Nov 29 16:09:04 disco-zfstest zed[2968]: Processing events since eid=0

  
  People in general would want the number of default running services low.
  Further than the general "not too much default daemons" there are a few 
issues like [1] which makes this worse if one is affected.

  There is no related upstream feature yet that I'd know of, nor did I
  find a trivial way to e.g. match a systemd unit Condition (which also
  is only evaluated only once, so it would miss later zfs creation) or a
  socket (what would it wait for).

  I wanted to ask here if you'd have more/better ideas or if we should
  directly file this upstream and then link that bug here?

  [1]: https://github.com/zfsonlinux/zfs/issues/7995

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1805982/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-02-27 Thread Randall Leeds
I have not yet had any problems with 2.0.2.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1805982] Re: Please consider making the zed daemon activated on demand

2021-02-27 Thread Randall Leeds
It seems like this may no longer be the case. In fact, a recent upgrade
on my system removed zfs-zed, which seems like maybe a mistake since the
zsys-setup script still seems to expect it.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1805982

Title:
  Please consider making the zed daemon activated on demand

Status in zfs-linux package in Ubuntu:
  New

Bug description:
  Hi,
  currently if one wants most common FS tools to be able to mount them to e.g. 
mount them "just in case" or when considering a more capable base image or 
install CD then there is the problem that installing "zfsutils-linux" pulls in 
"zfs-zed" as a recommends.

  Having it on disk would be ok, but IMHO there is no reason this daemon
  needs to run until the first ZFS element is created.

  ● zfs-zed.service - ZFS Event Daemon (zed)
 Loaded: loaded (/lib/systemd/system/zfs-zed.service; enabled; vendor 
preset: enabled)
 Active: active (running) since Thu 2018-11-29 16:09:04 UTC; 3min 26s ago
   Docs: man:zed(8)
   Main PID: 2968 (zed)
  Tasks: 2 (limit: 546)
 Memory: 436.0K
 CGroup: /system.slice/zfs-zed.service
 └─2968 /usr/sbin/zed -F

  Nov 29 16:09:04 disco-zfstest systemd[1]: Started ZFS Event Daemon (zed).
  Nov 29 16:09:04 disco-zfstest zed[2968]: ZFS Event Daemon 0.7.9-3ubuntu6 (PID 
2968)
  Nov 29 16:09:04 disco-zfstest zed[2968]: Processing events since eid=0

  
  People in general would want the number of default running services low.
  Further than the general "not too much default daemons" there are a few 
issues like [1] which makes this worse if one is affected.

  There is no related upstream feature yet that I'd know of, nor did I
  find a trivial way to e.g. match a systemd unit Condition (which also
  is only evaluated only once, so it would miss later zfs creation) or a
  socket (what would it wait for).

  I wanted to ask here if you'd have more/better ideas or if we should
  directly file this upstream and then link that bug here?

  [1]: https://github.com/zfsonlinux/zfs/issues/7995

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1805982/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-02-21 Thread Randall Leeds
I've install zfs-dkms to ensure that I'm using the 2.0.2 kmod. I'll post
back later in the week about whether I see any panics with regular use.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-02-21 Thread Randall Leeds
I'm trying out the ZFS 2.0.2 userspace packages now, from hursute (not
-proposed), and will report back. I note that the kernel module for
linux-modules-5.10.0-14-lowlatency that I'm using is still 2.0.1.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1906476] Re: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed

2021-02-15 Thread Randall Leeds
2.0.2 is now in hirsute-proposed.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1906476

Title:
  PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 ==
  sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED,
  &zp->z_sa_hdl)) failed

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS
  panic in the kernel log which was also hanging Disk I/O for all
  Chrome/Electron Apps.

  I have narrowed down a few important notes:
  - It does not happen with module version 0.8.4-1ubuntu11 built and included 
with 5.8.0-29-generic

  - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS
  on the same kernel and also on 5.8.18-acso (a custom kernel).

  - For whatever reason multiple Chrome/Electron apps were affected,
  specifically Discord, Chrome and Mattermost. In all cases they seem
  (but I was unable to strace the processes so it was a bit hard ot
  confirm 100% but by deduction from /proc/PID/fd and the hanging ls)
  they seem hung trying to open files in their 'Cache' directory, e.g.
  ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache ..
  while the issue was going on I could not list that directory either
  "ls" would just hang.

  - Once I removed zfs-dkms only to revert to the kernel built-in
  version it immediately worked without changing anything, removing
  files, etc.

  - It happened over multiple reboots and kernels every time, all my
  Chrome apps weren't working but for whatever reason nothing else
  seemed affected.

  - It would log a series of spl_panic dumps into kern.log that look like this:
  Dec  2 12:36:42 optane kernel: [   72.857033] VERIFY(0 == 
sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) 
failed
  Dec  2 12:36:42 optane kernel: [   72.857036] PANIC at 
zfs_znode.c:335:zfs_znode_sa_init()

  I could only find one other google reference to this issue, with 2 other 
users reporting the same error but on 20.04 here:
  https://github.com/openzfs/zfs/issues/10971

  - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure
  it was working on 0.8.4-1ubuntu15 but broken after upgrade to
  0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify
  that.

  There were a few originating call stacks but the first one I hit was

  Call Trace:
   dump_stack+0x74/0x95
   spl_dumpstack+0x29/0x2b [spl]
   spl_panic+0xd4/0xfc [spl]
   ? sa_cache_constructor+0x27/0x50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_buf_set_user_ie+0x54/0x80 [zfs]
   zfs_znode_sa_init+0xe0/0xf0 [zfs]
   zfs_znode_alloc+0x101/0x700 [zfs]
   ? arc_buf_fill+0x270/0xd30 [zfs]
   ? __cv_init+0x42/0x60 [spl]
   ? dnode_cons+0x28f/0x2a0 [zfs]
   ? _cond_resched+0x19/0x40
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? aggsum_add+0x153/0x170 [zfs]
   ? spl_kmem_alloc_impl+0xd8/0x110 [spl]
   ? arc_space_consume+0x54/0xe0 [zfs]
   ? dbuf_read+0x4a0/0xb50 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dnode_rele_and_unlock+0x5a/0xc0 [zfs]
   ? _cond_resched+0x19/0x40
   ? mutex_lock+0x12/0x40
   ? dmu_object_info_from_dnode+0x84/0xb0 [zfs]
   zfs_zget+0x1c3/0x270 [zfs]
   ? dmu_buf_rele+0x3a/0x40 [zfs]
   zfs_dirent_lock+0x349/0x680 [zfs]
   zfs_dirlook+0x90/0x2a0 [zfs]
   ? zfs_zaccess+0x10c/0x480 [zfs]
   zfs_lookup+0x202/0x3b0 [zfs]
   zpl_lookup+0xca/0x1e0 [zfs]
   path_openat+0x6a2/0xfe0
   do_filp_open+0x9b/0x110
   ? __check_object_size+0xdb/0x1b0
   ? __alloc_fd+0x46/0x170
   do_sys_openat2+0x217/0x2d0
   ? do_sys_openat2+0x217/0x2d0
   do_sys_open+0x59/0x80
   __x64_sys_openat+0x20/0x30

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1881701] Re: NVIDIA cards should have set AllowExternalGpus in /usr/share/X11/xorg.conf.d/10-nvidia.conf

2020-09-05 Thread Randall Leeds
** Also affects: nvidia-graphics-drivers-450 (Ubuntu)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to nvidia-graphics-drivers-440 in Ubuntu.
https://bugs.launchpad.net/bugs/1881701

Title:
  NVIDIA cards should have set AllowExternalGpus in
  /usr/share/X11/xorg.conf.d/10-nvidia.conf

Status in nvidia-graphics-drivers-440 package in Ubuntu:
  Confirmed
Status in nvidia-graphics-drivers-450 package in Ubuntu:
  New

Bug description:
  I have a Lenovo Thunderbold 3 Graphics Dock (comes integrated NVIDIA
  MTX 1050 graphics card), see
  https://support.lenovo.com/de/en/accessories/acc500012

  *What happened:*

  With nvidia-graphics-drivers-440 I just got a black screen. After
  looink into the /var/log/Xorg.0.log I noticed:

  ```

  (WW) NVIDIA(GPU-0): This device is an external GPU, but external
  GPUs have not

  (WW) NVIDIA(GPU-0): been enabled with AllowExternalGpus. Disabling
  this device

  (WW) NVIDIA(GPU-0): to prevent crashes from accidental removal.

  (EE) NVIDIA(0): Failing initialization of X screen 0
  ```

  Some googling turned up https://forums.developer.nvidia.com/t/egpu-
  and-intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4

  After I added the `Option "AllowExternalGpus" "true"` to `Section
  "OutputClass"` of `/usr/share/X11/xorg.conf.d/10-nvidia.conf` and
  rebooting the card worked.

  This is a matter for all eGPUs (external GPUs) that run NVIDIA I
  guess. See reports at
  
https://www.reddit.com/r/thinkpad/comments/fcuhkn/running_thunderbold_3_graphics_dock_nvidia_mtx/
  and aforementioned https://forums.developer.nvidia.com/t/egpu-and-
  intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4

  *Possible solution*

  Add this line to /usr/share/X11/xorg.conf.d/10-nvidia.conf:
  ```
  Section "OutputClass"
  Identifier "nvidia"
  MatchDriver "nvidia-drm"
  Driver "nvidia"
  Option "AllowEmptyInitialConfiguration"
  # v ADD THIS LINE v
  Option "AllowExternalGpus" "true"
  # ^ ADD THIS LINE ^^^
  ModulePath "/usr/lib/x86_64-linux-gnu/nvidia/xorg"
  EndSection
  ```

  ProblemType: Bug
  DistroRelease: Ubuntu 20.04
  Package: xserver-xorg-video-nvidia-440 440.82+really.440.64-0ubuntu6 
[modified: usr/share/X11/xorg.conf.d/10-nvidia.conf]
  ProcVersionSignature: Ubuntu 5.4.0-33.37-generic 5.4.34
  Uname: Linux 5.4.0-33-generic x86_64
  NonfreeKernelModules: nvidia_modeset nvidia
  ApportVersion: 2.20.11-0ubuntu27.2
  Architecture: amd64
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Date: Tue Jun  2 07:30:47 2020
  InstallationDate: Installed on 2019-09-19 (257 days ago)
  InstallationMedia: Ubuntu 19.04 "Disco Dingo" - Release amd64 (20190416)
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=de_DE.UTF-8
   SHELL=/bin/bash
  SourcePackage: nvidia-graphics-drivers-440
  UpgradeStatus: Upgraded to focal on 2020-05-29 (3 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-drivers-440/+bug/1881701/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1872108] Re: Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal

2020-08-22 Thread Randall Leeds
Benjamin Poirier seems to be working to upstream changes that would fix
this issue.

See these links for details:
https://github.com/gobenji/thinkpad-x1-gen7-sound
https://lore.kernel.org/alsa-devel/20200703080005.8942-1-benjamin.poir...@gmail.com/
https://lore.kernel.org/alsa-devel/20200703072302.16876-1-benjamin.poir...@gmail.com/

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to alsa-ucm-conf in Ubuntu.
https://bugs.launchpad.net/bugs/1872108

Title:
  Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal

Status in alsa-ucm-conf package in Ubuntu:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  Related to
  https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1845677/comments/14

  Audio works and the volume seems reasonable however the audio is
  noticeably missing bass giving everything a high treble sound.

  ProblemType: Bug
  DistroRelease: Ubuntu 20.04
  Package: alsa-ucm-conf 1.2.2-1
  ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27
  Uname: Linux 5.4.0-21-generic x86_64
  ApportVersion: 2.20.11-0ubuntu26
  Architecture: amd64
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Date: Fri Apr 10 09:17:33 2020
  Dependencies:

  InstallationDate: Installed on 2020-04-08 (1 days ago)
  InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402)
  PackageArchitecture: all
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  SourcePackage: alsa-ucm-conf
  UpgradeStatus: No upgrade log present (probably fresh install)
  --- 
  ProblemType: Bug
  ApportVersion: 2.20.11-0ubuntu26
  Architecture: amd64
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Dependencies:
   
  DistroRelease: Ubuntu 20.04
  InstallationDate: Installed on 2020-04-08 (1 days ago)
  InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402)
  Package: linux
  PackageArchitecture: all
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27
  Tags:  focal
  Uname: Linux 5.4.0-21-generic x86_64
  UpgradeStatus: No upgrade log present (probably fresh install)
  UserGroups: adm cdrom dip lpadmin lxd plugdev sambashare sudo
  _MarkForUpload: True

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/alsa-ucm-conf/+bug/1872108/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1872108] Re: Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal

2020-07-18 Thread Randall Leeds
This makes a world of difference. It seems to make the Headphone volume
control the bottom speakers while the Speaker volume still controls the
top speaker.

Does anyone know enough about ALSA to know if there's a way the command
could be tweaked so that the Speaker volume controls both?

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to alsa-ucm-conf in Ubuntu.
https://bugs.launchpad.net/bugs/1872108

Title:
  Bass not on working Lenovo ThinkPad X1 Carbon 7th gen running focal

Status in alsa-ucm-conf package in Ubuntu:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  Related to
  https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1845677/comments/14

  Audio works and the volume seems reasonable however the audio is
  noticeably missing bass giving everything a high treble sound.

  ProblemType: Bug
  DistroRelease: Ubuntu 20.04
  Package: alsa-ucm-conf 1.2.2-1
  ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27
  Uname: Linux 5.4.0-21-generic x86_64
  ApportVersion: 2.20.11-0ubuntu26
  Architecture: amd64
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Date: Fri Apr 10 09:17:33 2020
  Dependencies:

  InstallationDate: Installed on 2020-04-08 (1 days ago)
  InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402)
  PackageArchitecture: all
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  SourcePackage: alsa-ucm-conf
  UpgradeStatus: No upgrade log present (probably fresh install)
  --- 
  ProblemType: Bug
  ApportVersion: 2.20.11-0ubuntu26
  Architecture: amd64
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Dependencies:
   
  DistroRelease: Ubuntu 20.04
  InstallationDate: Installed on 2020-04-08 (1 days ago)
  InstallationMedia: Ubuntu 20.04 LTS "Focal Fossa" - Beta amd64 (20200402)
  Package: linux
  PackageArchitecture: all
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  ProcVersionSignature: Ubuntu 5.4.0-21.25-generic 5.4.27
  Tags:  focal
  Uname: Linux 5.4.0-21-generic x86_64
  UpgradeStatus: No upgrade log present (probably fresh install)
  UserGroups: adm cdrom dip lpadmin lxd plugdev sambashare sudo
  _MarkForUpload: True

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/alsa-ucm-conf/+bug/1872108/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1790515] Re: black screen on ubuntu GTX 1050 eGPU

2020-07-05 Thread Randall Leeds
Could you try the fix from #1881701, please?

https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-
drivers-440/+bug/1881701

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to nvidia-graphics-drivers-390 in Ubuntu.
https://bugs.launchpad.net/bugs/1790515

Title:
  black screen on ubuntu GTX 1050 eGPU

Status in nvidia-graphics-drivers-390 package in Ubuntu:
  New

Bug description:
  I'm using a laptop lenovo thinkpad T410 with intel graphics and egpu + GTX 
1050. (using expresscard or mini pcie both failed)
  When I install the 390 driver my graphical display fails. I know that egpu is 
not the best solution but for some people it works.

  I tried:
  390, 396, 384 nvidia drivers from ppa and from NVIDIA (.run)
  Ubuntu and Xubuntu (18, 17, 16) using gdm3 and lightgdm
  blacklist nouveau and everything similar

  Sometimes Ubuntu 18.04 freeze using live usb.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-drivers-390/+bug/1790515/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1881701] Re: NVIDIA cards should have set AllowExternalGpus in /usr/share/X11/xorg.conf.d/10-nvidia.conf

2020-07-05 Thread Randall Leeds
I can confirm that adding:

  Option "AllowExternalGpus"

Fixes my issue. This may be the same as #1790515, too.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to nvidia-graphics-drivers-440 in Ubuntu.
https://bugs.launchpad.net/bugs/1881701

Title:
  NVIDIA cards should have set AllowExternalGpus in
  /usr/share/X11/xorg.conf.d/10-nvidia.conf

Status in nvidia-graphics-drivers-440 package in Ubuntu:
  New

Bug description:
  I have a Lenovo Thunderbold 3 Graphics Dock (comes integrated NVIDIA
  MTX 1050 graphics card), see
  https://support.lenovo.com/de/en/accessories/acc500012

  *What happened:*

  With nvidia-graphics-drivers-440 I just got a black screen. After
  looink into the /var/log/Xorg.0.log I noticed:

  ```

  (WW) NVIDIA(GPU-0): This device is an external GPU, but external
  GPUs have not

  (WW) NVIDIA(GPU-0): been enabled with AllowExternalGpus. Disabling
  this device

  (WW) NVIDIA(GPU-0): to prevent crashes from accidental removal.

  (EE) NVIDIA(0): Failing initialization of X screen 0
  ```

  Some googling turned up https://forums.developer.nvidia.com/t/egpu-
  and-intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4

  After I added the `Option "AllowExternalGpus" "true"` to `Section
  "OutputClass"` of `/usr/share/X11/xorg.conf.d/10-nvidia.conf` and
  rebooting the card worked.

  This is a matter for all eGPUs (external GPUs) that run NVIDIA I
  guess. See reports at
  
https://www.reddit.com/r/thinkpad/comments/fcuhkn/running_thunderbold_3_graphics_dock_nvidia_mtx/
  and aforementioned https://forums.developer.nvidia.com/t/egpu-and-
  intel-nuc-xorg-conf-issues-login-loop-and-hanging/69409/4

  *Possible solution*

  Add this line to /usr/share/X11/xorg.conf.d/10-nvidia.conf:
  ```
  Section "OutputClass"
  Identifier "nvidia"
  MatchDriver "nvidia-drm"
  Driver "nvidia"
  Option "AllowEmptyInitialConfiguration"
  # v ADD THIS LINE v
  Option "AllowExternalGpus" "true"
  # ^ ADD THIS LINE ^^^
  ModulePath "/usr/lib/x86_64-linux-gnu/nvidia/xorg"
  EndSection
  ```

  ProblemType: Bug
  DistroRelease: Ubuntu 20.04
  Package: xserver-xorg-video-nvidia-440 440.82+really.440.64-0ubuntu6 
[modified: usr/share/X11/xorg.conf.d/10-nvidia.conf]
  ProcVersionSignature: Ubuntu 5.4.0-33.37-generic 5.4.34
  Uname: Linux 5.4.0-33-generic x86_64
  NonfreeKernelModules: nvidia_modeset nvidia
  ApportVersion: 2.20.11-0ubuntu27.2
  Architecture: amd64
  CasperMD5CheckResult: skip
  CurrentDesktop: ubuntu:GNOME
  Date: Tue Jun  2 07:30:47 2020
  InstallationDate: Installed on 2019-09-19 (257 days ago)
  InstallationMedia: Ubuntu 19.04 "Disco Dingo" - Release amd64 (20190416)
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=de_DE.UTF-8
   SHELL=/bin/bash
  SourcePackage: nvidia-graphics-drivers-440
  UpgradeStatus: Upgraded to focal on 2020-05-29 (3 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/nvidia-graphics-drivers-440/+bug/1881701/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root

2017-03-20 Thread Randall Leeds
Confirmed fixed over here. Thanks for the fast turnaround!

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1673197

Title:
  Upgrading zfs-initramfs breaks booting from a zfs root

Status in zfs-linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Debian:
  Fix Released

Bug description:
  Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot.

  I've EFI and a small ext2 boot partition, and the rest of my SDD is a
  zfs pool. After upgrade I'm dropped in the initramfs shell and no
  zpools are available.

  Downgrading to 0.6.5.8-0ubuntu9 makes it boot again.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.04
  Package: zfs-initramfs 0.6.5.9-4ubuntu1
  ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1
  Uname: Linux 4.10.0-11-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.4-0ubuntu2
  Architecture: amd64
  CurrentDesktop: Unity:Unity7
  Date: Wed Mar 15 18:29:04 2017
  InstallationDate: Installed on 2016-05-06 (312 days ago)
  InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 
(20160420.1)
  PackageArchitecture: all
  SourcePackage: zfs-linux
  UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root

2017-03-16 Thread Randall Leeds
My last comment got cut off. I was going to say I don't see /usr/share
/initramfs-tools/conf.d/zfs on my filesystem.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1673197

Title:
  Upgrading zfs-initramfs breaks booting from a zfs root

Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot.

  I've EFI and a small ext2 boot partition, and the rest of my SDD is a
  zfs pool. After upgrade I'm dropped in the initramfs shell and no
  zpools are available.

  Downgrading to 0.6.5.8-0ubuntu9 makes it boot again.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.04
  Package: zfs-initramfs 0.6.5.9-4ubuntu1
  ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1
  Uname: Linux 4.10.0-11-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.4-0ubuntu2
  Architecture: amd64
  CurrentDesktop: Unity:Unity7
  Date: Wed Mar 15 18:29:04 2017
  InstallationDate: Installed on 2016-05-06 (312 days ago)
  InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 
(20160420.1)
  PackageArchitecture: all
  SourcePackage: zfs-linux
  UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root

2017-03-16 Thread Randall Leeds
Maybe this is some regression in zfs-initramfs?

I see some mention of this in 
https://bugs.launchpad.net/ubuntu/+source/zfs-initramfs/+bug/1530953
I don

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1673197

Title:
  Upgrading zfs-initramfs breaks booting from a zfs root

Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot.

  I've EFI and a small ext2 boot partition, and the rest of my SDD is a
  zfs pool. After upgrade I'm dropped in the initramfs shell and no
  zpools are available.

  Downgrading to 0.6.5.8-0ubuntu9 makes it boot again.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.04
  Package: zfs-initramfs 0.6.5.9-4ubuntu1
  ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1
  Uname: Linux 4.10.0-11-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.4-0ubuntu2
  Architecture: amd64
  CurrentDesktop: Unity:Unity7
  Date: Wed Mar 15 18:29:04 2017
  InstallationDate: Installed on 2016-05-06 (312 days ago)
  InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 
(20160420.1)
  PackageArchitecture: all
  SourcePackage: zfs-linux
  UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root

2017-03-16 Thread Randall Leeds
I just happened to find that parameter mentioned in some ZFS root setup
guides and tried it as a wild guess.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1673197

Title:
  Upgrading zfs-initramfs breaks booting from a zfs root

Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot.

  I've EFI and a small ext2 boot partition, and the rest of my SDD is a
  zfs pool. After upgrade I'm dropped in the initramfs shell and no
  zpools are available.

  Downgrading to 0.6.5.8-0ubuntu9 makes it boot again.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.04
  Package: zfs-initramfs 0.6.5.9-4ubuntu1
  ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1
  Uname: Linux 4.10.0-11-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.4-0ubuntu2
  Architecture: amd64
  CurrentDesktop: Unity:Unity7
  Date: Wed Mar 15 18:29:04 2017
  InstallationDate: Installed on 2016-05-06 (312 days ago)
  InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 
(20160420.1)
  PackageArchitecture: all
  SourcePackage: zfs-linux
  UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1673197] Re: Upgrading zfs-initramfs breaks booting from a zfs root

2017-03-16 Thread Randall Leeds
I don't know what changed, but adding `boot=zfs` to my grub kernel
parameters fixes this. I never needed it in the past and `update-grub`
doesn't create it.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1673197

Title:
  Upgrading zfs-initramfs breaks booting from a zfs root

Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Upgrading to zfs-initramfs 0.6.5.9-4ubuntu1 breaks my boot.

  I've EFI and a small ext2 boot partition, and the rest of my SDD is a
  zfs pool. After upgrade I'm dropped in the initramfs shell and no
  zpools are available.

  Downgrading to 0.6.5.8-0ubuntu9 makes it boot again.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.04
  Package: zfs-initramfs 0.6.5.9-4ubuntu1
  ProcVersionSignature: Ubuntu 4.10.0-11.13-generic 4.10.1
  Uname: Linux 4.10.0-11-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.4-0ubuntu2
  Architecture: amd64
  CurrentDesktop: Unity:Unity7
  Date: Wed Mar 15 18:29:04 2017
  InstallationDate: Installed on 2016-05-06 (312 days ago)
  InstallationMedia: Ubuntu 16.04 LTS "Xenial Xerus" - Release amd64 
(20160420.1)
  PackageArchitecture: all
  SourcePackage: zfs-linux
  UpgradeStatus: Upgraded to zesty on 2016-11-22 (113 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1673197/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1289797] Re: [LENOVO 20A8A01DLM] suspend/resume failure

2014-04-13 Thread Randall Leeds
Confirm fixed by the most recent BIOS update.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1289797

Title:
  [LENOVO 20A8A01DLM] suspend/resume failure

Status in “linux” package in Ubuntu:
  Triaged

Bug description:
  It dies when I suspend it, the adaptive key strip and the power key
  keep turned on.

  ProblemType: KernelOops
  DistroRelease: Ubuntu 14.04
  Package: linux-image-3.13.0-16-generic 3.13.0-16.36
  ProcVersionSignature: Ubuntu 3.13.0-16.36-generic 3.13.5
  Uname: Linux 3.13.0-16-generic i686
  Annotation: This occured during a previous suspend and prevented it from 
resuming properly.
  ApportVersion: 2.13.2-0ubuntu5
  Architecture: i386
  AudioDevicesInUse:
   USERPID ACCESS COMMAND
   /dev/snd/controlC0:  ubuntu 3349 F pulseaudio
   /dev/snd/controlC1:  ubuntu 3349 F pulseaudio
  CRDA:
   country MX:
(2402 - 2472 @ 40), (3, 27)
(5170 - 5250 @ 40), (3, 17)
(5250 - 5330 @ 40), (3, 23), DFS
(5735 - 5835 @ 40), (3, 30)
  CasperVersion: 1.337
  Date: Fri Mar  7 21:10:38 2014
  ExecutablePath: /usr/share/apport/apportcheckresume
  Failure: suspend/resume
  InterpreterPath: /usr/bin/python3.4
  LiveMediaBuild: Ubuntu 14.04 LTS "Trusty Tahr" - Alpha i386 (20140307)
  MachineType: LENOVO 20A8A01DLM
  ProcCmdline: /usr/bin/python3 /usr/share/apport/apportcheckresume
  ProcEnviron:
   TERM=linux
   PATH=(custom, no user)
  ProcFB: 0 inteldrmfb
  ProcKernelCmdLine: noprompt cdrom-detect/try-usb=true persistent 
file=/cdrom/preseed/hostname.seed boot=casper initrd=/casper/initrd.lz quiet 
splash -- maybe-ubiquity
  PulseList: Error: command ['pacmd', 'list'] failed with exit code 1: No 
PulseAudio daemon running, or not running as session daemon.
  RelatedPackageVersions:
   linux-restricted-modules-3.13.0-16-generic N/A
   linux-backports-modules-3.13.0-16-generic  N/A
   linux-firmware 1.126
  SourcePackage: linux
  Title: [LENOVO 20A8A01DLM] suspend/resume failure
  UpgradeStatus: No upgrade log present (probably fresh install)
  UserGroups:
   
  dmi.bios.date: 12/03/2013
  dmi.bios.vendor: LENOVO
  dmi.bios.version: GRET29WW (1.06 )
  dmi.board.asset.tag: Not Available
  dmi.board.name: 20A8A01DLM
  dmi.board.vendor: LENOVO
  dmi.board.version: SDK0E50510 Pro
  dmi.chassis.asset.tag: No Asset Information
  dmi.chassis.type: 10
  dmi.chassis.vendor: LENOVO
  dmi.chassis.version: Not Available
  dmi.modalias: 
dmi:bvnLENOVO:bvrGRET29WW(1.06):bd12/03/2013:svnLENOVO:pn20A8A01DLM:pvrThinkPadX1Carbon2nd:rvnLENOVO:rn20A8A01DLM:rvrSDK0E50510Pro:cvnLENOVO:ct10:cvrNotAvailable:
  dmi.product.name: 20A8A01DLM
  dmi.product.version: ThinkPad X1 Carbon 2nd
  dmi.sys.vendor: LENOVO

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1289797/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp