[Kernel-packages] [Bug 1869372] [NEW] Popping/crackling noise when sound card wakes from powersaving

2020-03-27 Thread Simos Xenitellis
Public bug reported:

I could not find an existing report on this.
When Ubuntu 20.04 does powersaving, the audio card powers down.
When the card powers up again, it emits a short annoying popping/crackling 
noise, then audio plays just fine.

The workaround is to disable power-saving for the audio card, by adding

options snd_hda_intel power_save=0 power_save_controller=N

to a file in /etc/modprobe.d/alsa-intel-no-powersave.conf

and then reboot.

** Affects: linux (Ubuntu)
 Importance: Undecided
 Status: New

** Description changed:

  I could not find an existing report on this.
- When Ubuntu does powersaving, the audio card powers down.
+ When Ubuntu 20.04 does powersaving, the audio card powers down.
  When the card powers up again, it emits a short annoying popping/crackling 
noise, then audio plays just fine.
  
  The workaround is to disable power-saving for the audio card, by adding
  
  options snd_hda_intel power_save=0 power_save_controller=N
  
  to a file in /etc/modprobe.d/alsa-intel-no-powersave.conf
  
  and then reboot.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1869372

Title:
  Popping/crackling noise when sound card wakes from powersaving

Status in linux package in Ubuntu:
  New

Bug description:
  I could not find an existing report on this.
  When Ubuntu 20.04 does powersaving, the audio card powers down.
  When the card powers up again, it emits a short annoying popping/crackling 
noise, then audio plays just fine.

  The workaround is to disable power-saving for the audio card, by
  adding

  options snd_hda_intel power_save=0 power_save_controller=N

  to a file in /etc/modprobe.d/alsa-intel-no-powersave.conf

  and then reboot.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1869372/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1781601] Re: Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy css_killed_work_fn)

2018-10-03 Thread Simos Xenitellis
The system is largely unresponsive. I got this though:

ubuntu@myserver:~$ free
  totalusedfree  shared  buff/cache   available
Mem:   6568594848896836  803588  4691721598552415572340
Swap:   1996796   51712 1945084
ubuntu@myserver:~$ 

It appears that the memory has been exhausted.
Do kernel errors from resource exhaustion count as bugs?

I am changing from CONFIRMED to NEW.

** Changed in: linux (Ubuntu)
   Status: Confirmed => New

** Changed in: linux (Ubuntu Bionic)
   Status: Confirmed => New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781601

Title:
  Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy
  css_killed_work_fn)

Status in linux package in Ubuntu:
  New
Status in linux source package in Bionic:
  New

Bug description:
  I have been stress-testing LXD and managed to cause the following
  kernel hung:

  [10271.564074] INFO: task systemd:1 blocked for more than 120 seconds.
  [10271.570397]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.577212] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.585071] systemd D0 1  0 0x
  [10271.585073] Call Trace:
  [10271.585080]  __schedule+0x297/0x8b0
  [10271.585086]  schedule+0x2c/0x80
  [10271.585090]  schedule_preempt_disabled+0xe/0x10
  [10271.585095]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.585100]  __mutex_lock_slowpath+0x13/0x20
  [10271.585101]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.585102]  mutex_lock+0x2f/0x40
  [10271.585106]  proc_cgroup_show+0x4c/0x2a0
  [10271.585108]  proc_single_show+0x56/0x80
  [10271.585111]  seq_read+0xe5/0x430
  [10271.585114]  __vfs_read+0x1b/0x40
  [10271.585115]  vfs_read+0x8e/0x130
  [10271.585117]  SyS_read+0x55/0xc0
  [10271.585120]  do_syscall_64+0x73/0x130
  [10271.585121]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [10271.585123] RIP: 0033:0x7fd47634d0b4
  [10271.585124] RSP: 002b:7ffe9a186700 EFLAGS: 0246 ORIG_RAX: 

  [10271.585125] RAX: ffda RBX: 0021 RCX: 
7fd47634d0b4
  [10271.585126] RDX: 0400 RSI: 55f70a08c0c0 RDI: 
0021
  [10271.585127] RBP: 55f70a08c0c0 R08:  R09: 

  [10271.585128] R10:  R11: 0246 R12: 
0400
  [10271.585129] R13: 7fd4766252a0 R14: 55f70a0298e0 R15: 
07ff
  [10271.585221] INFO: task lxcfs:84510 blocked for more than 120 seconds.
  [10271.591687]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.598531] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.606401] lxcfs   D0 84510  1 0x
  [10271.606407] Call Trace:
  [10271.606416]  __schedule+0x297/0x8b0
  [10271.606418]  schedule+0x2c/0x80
  [10271.606420]  schedule_preempt_disabled+0xe/0x10
  [10271.606421]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.606423]  __mutex_lock_slowpath+0x13/0x20
  [10271.606424]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.606425]  mutex_lock+0x2f/0x40
  [10271.606427]  proc_cgroup_show+0x4c/0x2a0
  [10271.606429]  proc_single_show+0x56/0x80
  [10271.606432]  seq_read+0xe5/0x430
  [10271.606434]  __vfs_read+0x1b/0x40
  [10271.606436]  vfs_read+0x8e/0x130
  [10271.606437]  SyS_read+0x55/0xc0
  [10271.606440]  do_syscall_64+0x73/0x130
  [10271.606441]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [10271.606443] RIP: 0033:0x7f1a50e700b4
  [10271.606443] RSP: 002b:7f1a2b7fd870 EFLAGS: 0246 ORIG_RAX: 

  [10271.606445] RAX: ffda RBX: 0013 RCX: 
7f1a50e700b4
  [10271.606446] RDX: 0400 RSI: 7f1a04000f90 RDI: 
0013
  [10271.606446] RBP: 7f1a04000f90 R08: 0001 R09: 

  [10271.606447] R10:  R11: 0246 R12: 
0400
  [10271.606448] R13: 7f1a511482a0 R14:  R15: 
7f1a38003a40
  [10271.606450] INFO: task lxcfs:84676 blocked for more than 120 seconds.
  [10271.612911]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.619722] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.627577] lxcfs   D0 84676  1 0x
  [10271.627579] Call Trace:
  [10271.627583]  __schedule+0x297/0x8b0
  [10271.627589]  schedule+0x2c/0x80
  [10271.627594]  schedule_preempt_disabled+0xe/0x10
  [10271.627598]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.627600]  __mutex_lock_slowpath+0x13/0x20
  [10271.627601]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.627602]  mutex_lock+0x2f/0x40
  [10271.627604]  proc_cgroup_show+0x4c/0x2a0
  [10271.627606]  proc_single_show+0x56/0x80
  [10271.627608]  seq_read+0xe5/0x430
  [10271.627610]  __vfs_read+0x1b/0x40
  [10271.627611]  

[Kernel-packages] [Bug 1781601] Re: Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy css_killed_work_fn)

2018-10-03 Thread Simos Xenitellis
I got the same bug again. Here are the kernel messages:


[ 1450.993972] INFO: task systemd:1 blocked for more than 120 seconds.
[ 1451.000279]   Tainted: P   O 4.15.0-36-generic #39-Ubuntu
[ 1451.007094] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[ 1451.014957] systemd D0 1  0 0x
[ 1451.014960] Call Trace:
[ 1451.014969]  __schedule+0x291/0x8a0
[ 1451.014971]  schedule+0x2c/0x80
[ 1451.014973]  schedule_preempt_disabled+0xe/0x10
[ 1451.014974]  __mutex_lock.isra.2+0x18c/0x4d0
[ 1451.014976]  __mutex_lock_slowpath+0x13/0x20
[ 1451.014978]  ? __mutex_lock_slowpath+0x13/0x20
[ 1451.014979]  mutex_lock+0x2f/0x40
[ 1451.014982]  proc_cgroup_show+0x4c/0x2a0
[ 1451.014985]  proc_single_show+0x56/0x80
[ 1451.014988]  seq_read+0xe5/0x430
[ 1451.014990]  __vfs_read+0x1b/0x40
[ 1451.014991]  vfs_read+0x8e/0x130
[ 1451.014992]  SyS_read+0x55/0xc0
[ 1451.014995]  do_syscall_64+0x73/0x130
[ 1451.014997]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[ 1451.014999] RIP: 0033:0x7fc9a5300081
[ 1451.015000] RSP: 002b:7ffcdf16ab48 EFLAGS: 0246 ORIG_RAX: 

[ 1451.015002] RAX: ffda RBX: 55c5a2612290 RCX: 7fc9a5300081
[ 1451.015003] RDX: 0400 RSI: 55c5a269b4d0 RDI: 0026
[ 1451.015004] RBP: 0d68 R08: 0001 R09: 
[ 1451.015004] R10:  R11: 0246 R12: 7fc9a55d7760
[ 1451.015005] R13: 7fc9a55d82a0 R14: 55c5a2612290 R15: 07ff
[ 1451.015077] INFO: task systemd-journal:811 blocked for more than 120 seconds.
[ 1451.022239]   Tainted: P   O 4.15.0-36-generic #39-Ubuntu
[ 1451.029073] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[ 1451.036938] systemd-journal D0   811  1 0x0120
[ 1451.036942] Call Trace:
[ 1451.036950]  __schedule+0x291/0x8a0
[ 1451.036954]  ? ___slab_alloc+0x20a/0x4b0
[ 1451.036956]  schedule+0x2c/0x80
[ 1451.036957]  schedule_preempt_disabled+0xe/0x10
[ 1451.036958]  __mutex_lock.isra.2+0x18c/0x4d0
[ 1451.036960]  __mutex_lock_slowpath+0x13/0x20
[ 1451.036962]  ? __mutex_lock_slowpath+0x13/0x20
[ 1451.036963]  mutex_lock+0x2f/0x40
[ 1451.036966]  proc_cgroup_show+0x4c/0x2a0
[ 1451.036969]  proc_single_show+0x56/0x80
[ 1451.036972]  seq_read+0xe5/0x430
[ 1451.036975]  __vfs_read+0x1b/0x40
[ 1451.036978]  vfs_read+0x8e/0x130
[ 1451.036981]  SyS_read+0x55/0xc0
[ 1451.036985]  do_syscall_64+0x73/0x130
[ 1451.036988]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[ 1451.036991] RIP: 0033:0x7f331df97081
[ 1451.036993] RSP: 002b:7ffd719a04c8 EFLAGS: 0246 ORIG_RAX: 

[ 1451.036997] RAX: ffda RBX: 56046305a6c0 RCX: 7f331df97081
[ 1451.037000] RDX: 0400 RSI: 560463012a10 RDI: 001f
[ 1451.037001] RBP: 0d68 R08: 0001 R09: 
[ 1451.037003] R10:  R11: 0246 R12: 7f331e26e760
[ 1451.037005] R13: 7f331e26f2a0 R14: 56046305a6c0 R15: 07ff
[ 1451.037029] INFO: task lxcfs:39982 blocked for more than 120 seconds.
[ 1451.043498]   Tainted: P   O 4.15.0-36-generic #39-Ubuntu
[ 1451.050308] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[ 1451.058165] lxcfs   D0 39982  1 0x
[ 1451.058167] Call Trace:
[ 1451.058172]  __schedule+0x291/0x8a0
[ 1451.058174]  schedule+0x2c/0x80
[ 1451.058175]  schedule_preempt_disabled+0xe/0x10
[ 1451.058176]  __mutex_lock.isra.2+0x18c/0x4d0
[ 1451.058178]  __mutex_lock_slowpath+0x13/0x20
[ 1451.058179]  ? __mutex_lock_slowpath+0x13/0x20
[ 1451.058180]  mutex_lock+0x2f/0x40
[ 1451.058182]  proc_cgroup_show+0x4c/0x2a0
[ 1451.058184]  proc_single_show+0x56/0x80
[ 1451.058185]  seq_read+0xe5/0x430
[ 1451.058187]  __vfs_read+0x1b/0x40
[ 1451.058188]  vfs_read+0x8e/0x130
[ 1451.058189]  SyS_read+0x55/0xc0
[ 1451.058191]  do_syscall_64+0x73/0x130
[ 1451.058192]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[ 1451.058194] RIP: 0033:0x7fe0461c70b4
[ 1451.058194] RSP: 002b:7fe036ffc870 EFLAGS: 0246 ORIG_RAX: 

[ 1451.058196] RAX: ffda RBX: 001a RCX: 7fe0461c70b4
[ 1451.058196] RDX: 0400 RSI: 7fdfb00231e0 RDI: 001a
[ 1451.058197] RBP: 7fdfb00231e0 R08: 0001 R09: 
[ 1451.058198] R10:  R11: 0246 R12: 0400
[ 1451.058198] R13: 7fe04649f2a0 R14:  R15: 
[ 1451.058200] INFO: task lxcfs:118730 blocked for more than 120 seconds.
[ 1451.064746]   Tainted: P   O 4.15.0-36-generic #39-Ubuntu
[ 1451.071559] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[ 1451.079411] lxcfs   D0 118730  1 0x
[ 1451.079415] Call Trace:
[ 1451.079418]  __schedule+0x291/0x8a0
[ 1451.079421]  ? 

[Kernel-packages] [Bug 1781601] Re: Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy css_killed_work_fn)

2018-10-03 Thread Simos Xenitellis
I tried the following command:


ubuntu@myserver:~$ lxd-benchmark launch --count 900 --parallel 24 ubuntu:18.04
Test environment:
  Server backend: lxd
  Server version: 3.0.2
  Kernel: Linux
  Kernel architecture: x86_64
  Kernel version: 4.15.0-36-generic
  Storage backend: zfs
  Storage version: 0.7.5-1ubuntu16.3
  Container backend: lxc
  Container version: 3.0.2

Test variables:
  Container count: 900
  Container mode: unprivileged
  Startup mode: normal startup
  Image: ubuntu:18.04
  Batches: 37
  Batch size: 24
  Remainder: 12

[Oct  3 21:46:55.617] Found image in local store: 
c395a7105278712478ec1dbfaab1865593fc11292f99afe01d5b94f1c34a9a3a
[Oct  3 21:46:55.617] Batch processing start
[Oct  3 21:47:09.310] Processed 24 containers in 13.693s (1.753/s)
[Oct  3 21:47:26.739] Processed 48 containers in 31.122s (1.542/s)
[Oct  3 21:48:06.052] Processed 96 containers in 70.435s (1.363/s)
[Oct  3 21:49:27.340] Processed 192 containers in 151.723s (1.265/s)
^C

I interrupted the benchmark because it got stuck.

Note:

1. I am running Ubuntu 18.04.
2. With the updated kernel 4.15.0-36. I did not try the proposed kernel.
3. With LXD 3.0.2 (from bionic/proposed)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781601

Title:
  Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy
  css_killed_work_fn)

Status in linux package in Ubuntu:
  Confirmed
Status in linux source package in Bionic:
  Confirmed

Bug description:
  I have been stress-testing LXD and managed to cause the following
  kernel hung:

  [10271.564074] INFO: task systemd:1 blocked for more than 120 seconds.
  [10271.570397]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.577212] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.585071] systemd D0 1  0 0x
  [10271.585073] Call Trace:
  [10271.585080]  __schedule+0x297/0x8b0
  [10271.585086]  schedule+0x2c/0x80
  [10271.585090]  schedule_preempt_disabled+0xe/0x10
  [10271.585095]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.585100]  __mutex_lock_slowpath+0x13/0x20
  [10271.585101]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.585102]  mutex_lock+0x2f/0x40
  [10271.585106]  proc_cgroup_show+0x4c/0x2a0
  [10271.585108]  proc_single_show+0x56/0x80
  [10271.585111]  seq_read+0xe5/0x430
  [10271.585114]  __vfs_read+0x1b/0x40
  [10271.585115]  vfs_read+0x8e/0x130
  [10271.585117]  SyS_read+0x55/0xc0
  [10271.585120]  do_syscall_64+0x73/0x130
  [10271.585121]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [10271.585123] RIP: 0033:0x7fd47634d0b4
  [10271.585124] RSP: 002b:7ffe9a186700 EFLAGS: 0246 ORIG_RAX: 

  [10271.585125] RAX: ffda RBX: 0021 RCX: 
7fd47634d0b4
  [10271.585126] RDX: 0400 RSI: 55f70a08c0c0 RDI: 
0021
  [10271.585127] RBP: 55f70a08c0c0 R08:  R09: 

  [10271.585128] R10:  R11: 0246 R12: 
0400
  [10271.585129] R13: 7fd4766252a0 R14: 55f70a0298e0 R15: 
07ff
  [10271.585221] INFO: task lxcfs:84510 blocked for more than 120 seconds.
  [10271.591687]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.598531] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.606401] lxcfs   D0 84510  1 0x
  [10271.606407] Call Trace:
  [10271.606416]  __schedule+0x297/0x8b0
  [10271.606418]  schedule+0x2c/0x80
  [10271.606420]  schedule_preempt_disabled+0xe/0x10
  [10271.606421]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.606423]  __mutex_lock_slowpath+0x13/0x20
  [10271.606424]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.606425]  mutex_lock+0x2f/0x40
  [10271.606427]  proc_cgroup_show+0x4c/0x2a0
  [10271.606429]  proc_single_show+0x56/0x80
  [10271.606432]  seq_read+0xe5/0x430
  [10271.606434]  __vfs_read+0x1b/0x40
  [10271.606436]  vfs_read+0x8e/0x130
  [10271.606437]  SyS_read+0x55/0xc0
  [10271.606440]  do_syscall_64+0x73/0x130
  [10271.606441]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [10271.606443] RIP: 0033:0x7f1a50e700b4
  [10271.606443] RSP: 002b:7f1a2b7fd870 EFLAGS: 0246 ORIG_RAX: 

  [10271.606445] RAX: ffda RBX: 0013 RCX: 
7f1a50e700b4
  [10271.606446] RDX: 0400 RSI: 7f1a04000f90 RDI: 
0013
  [10271.606446] RBP: 7f1a04000f90 R08: 0001 R09: 

  [10271.606447] R10:  R11: 0246 R12: 
0400
  [10271.606448] R13: 7f1a511482a0 R14:  R15: 
7f1a38003a40
  [10271.606450] INFO: task lxcfs:84676 blocked for more than 120 seconds.
  [10271.612911]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.619722] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this 

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-07-18 Thread Simos Xenitellis
@Vasiliy: Indeed.

The version in -proposed is "4.15.0-29.31" (source:
https://launchpad.net/ubuntu/bionic/+queue?queue_state=3_text
=linux-image)

The page for that version at 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1782173
does not have a reference to this bug number #1781364.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Committed
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Confirmed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Confirmed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Committed
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch 

[Kernel-packages] [Bug 1781601] Re: Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy css_killed_work_fn)

2018-07-17 Thread Simos Xenitellis
I have been doing stress-testing on LXD at a freshly installed 18.04.
The Linux kernel was the standard 18.04 kernel.
LXD though was compiled from master, ZFS was also compile from master. 

I performed the stress-testing by running the command

lxd-benchmark --count 384 --parallel 24

This launches 384 Ubuntu 16.04 containers in batches of 24 containers.

LXD does not run well on the mainline Linux kernel because I think some 
necessary patches
have not been upstreamed yet. 

I plan to do more stress testing, and when I get the same issue,
I'll run apport to retrieve information from the system to attach here. 
Obviously, when I deploy the server, I'll prepare it for apport beforehand.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781601

Title:
  Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy
  css_killed_work_fn)

Status in linux package in Ubuntu:
  Incomplete
Status in linux source package in Bionic:
  Incomplete

Bug description:
  I have been stress-testing LXD and managed to cause the following
  kernel hung:

  [10271.564074] INFO: task systemd:1 blocked for more than 120 seconds.
  [10271.570397]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.577212] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.585071] systemd D0 1  0 0x
  [10271.585073] Call Trace:
  [10271.585080]  __schedule+0x297/0x8b0
  [10271.585086]  schedule+0x2c/0x80
  [10271.585090]  schedule_preempt_disabled+0xe/0x10
  [10271.585095]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.585100]  __mutex_lock_slowpath+0x13/0x20
  [10271.585101]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.585102]  mutex_lock+0x2f/0x40
  [10271.585106]  proc_cgroup_show+0x4c/0x2a0
  [10271.585108]  proc_single_show+0x56/0x80
  [10271.585111]  seq_read+0xe5/0x430
  [10271.585114]  __vfs_read+0x1b/0x40
  [10271.585115]  vfs_read+0x8e/0x130
  [10271.585117]  SyS_read+0x55/0xc0
  [10271.585120]  do_syscall_64+0x73/0x130
  [10271.585121]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [10271.585123] RIP: 0033:0x7fd47634d0b4
  [10271.585124] RSP: 002b:7ffe9a186700 EFLAGS: 0246 ORIG_RAX: 

  [10271.585125] RAX: ffda RBX: 0021 RCX: 
7fd47634d0b4
  [10271.585126] RDX: 0400 RSI: 55f70a08c0c0 RDI: 
0021
  [10271.585127] RBP: 55f70a08c0c0 R08:  R09: 

  [10271.585128] R10:  R11: 0246 R12: 
0400
  [10271.585129] R13: 7fd4766252a0 R14: 55f70a0298e0 R15: 
07ff
  [10271.585221] INFO: task lxcfs:84510 blocked for more than 120 seconds.
  [10271.591687]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.598531] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.606401] lxcfs   D0 84510  1 0x
  [10271.606407] Call Trace:
  [10271.606416]  __schedule+0x297/0x8b0
  [10271.606418]  schedule+0x2c/0x80
  [10271.606420]  schedule_preempt_disabled+0xe/0x10
  [10271.606421]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.606423]  __mutex_lock_slowpath+0x13/0x20
  [10271.606424]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.606425]  mutex_lock+0x2f/0x40
  [10271.606427]  proc_cgroup_show+0x4c/0x2a0
  [10271.606429]  proc_single_show+0x56/0x80
  [10271.606432]  seq_read+0xe5/0x430
  [10271.606434]  __vfs_read+0x1b/0x40
  [10271.606436]  vfs_read+0x8e/0x130
  [10271.606437]  SyS_read+0x55/0xc0
  [10271.606440]  do_syscall_64+0x73/0x130
  [10271.606441]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [10271.606443] RIP: 0033:0x7f1a50e700b4
  [10271.606443] RSP: 002b:7f1a2b7fd870 EFLAGS: 0246 ORIG_RAX: 

  [10271.606445] RAX: ffda RBX: 0013 RCX: 
7f1a50e700b4
  [10271.606446] RDX: 0400 RSI: 7f1a04000f90 RDI: 
0013
  [10271.606446] RBP: 7f1a04000f90 R08: 0001 R09: 

  [10271.606447] R10:  R11: 0246 R12: 
0400
  [10271.606448] R13: 7f1a511482a0 R14:  R15: 
7f1a38003a40
  [10271.606450] INFO: task lxcfs:84676 blocked for more than 120 seconds.
  [10271.612911]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
  [10271.619722] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [10271.627577] lxcfs   D0 84676  1 0x
  [10271.627579] Call Trace:
  [10271.627583]  __schedule+0x297/0x8b0
  [10271.627589]  schedule+0x2c/0x80
  [10271.627594]  schedule_preempt_disabled+0xe/0x10
  [10271.627598]  __mutex_lock.isra.2+0x18c/0x4d0
  [10271.627600]  __mutex_lock_slowpath+0x13/0x20
  [10271.627601]  ? __mutex_lock_slowpath+0x13/0x20
  [10271.627602]  mutex_lock+0x2f/0x40
  [10271.627604]  proc_cgroup_show+0x4c/0x2a0
  

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-07-16 Thread Simos Xenitellis
zfsutils-linux (zfs-linux, zfs-linux_0.7.5-1ubuntu16.3) is already in proposed,
https://launchpad.net/ubuntu/bionic/+queue?queue_state=3_text=zfs-linux

Please report when linux-image gets into -proposed,
https://launchpad.net/ubuntu/bionic/+queue?queue_state=3_text=linux-image

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Committed
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Confirmed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Confirmed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Committed
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 

[Kernel-packages] [Bug 1781601] [NEW] Stress-testing LXD causes kernel hung in cgroups (cgroup_destroy css_killed_work_fn)

2018-07-13 Thread Simos Xenitellis
Public bug reported:

I have been stress-testing LXD and managed to cause the following kernel
hung:

[10271.564074] INFO: task systemd:1 blocked for more than 120 seconds.
[10271.570397]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
[10271.577212] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[10271.585071] systemd D0 1  0 0x
[10271.585073] Call Trace:
[10271.585080]  __schedule+0x297/0x8b0
[10271.585086]  schedule+0x2c/0x80
[10271.585090]  schedule_preempt_disabled+0xe/0x10
[10271.585095]  __mutex_lock.isra.2+0x18c/0x4d0
[10271.585100]  __mutex_lock_slowpath+0x13/0x20
[10271.585101]  ? __mutex_lock_slowpath+0x13/0x20
[10271.585102]  mutex_lock+0x2f/0x40
[10271.585106]  proc_cgroup_show+0x4c/0x2a0
[10271.585108]  proc_single_show+0x56/0x80
[10271.585111]  seq_read+0xe5/0x430
[10271.585114]  __vfs_read+0x1b/0x40
[10271.585115]  vfs_read+0x8e/0x130
[10271.585117]  SyS_read+0x55/0xc0
[10271.585120]  do_syscall_64+0x73/0x130
[10271.585121]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[10271.585123] RIP: 0033:0x7fd47634d0b4
[10271.585124] RSP: 002b:7ffe9a186700 EFLAGS: 0246 ORIG_RAX: 

[10271.585125] RAX: ffda RBX: 0021 RCX: 7fd47634d0b4
[10271.585126] RDX: 0400 RSI: 55f70a08c0c0 RDI: 0021
[10271.585127] RBP: 55f70a08c0c0 R08:  R09: 
[10271.585128] R10:  R11: 0246 R12: 0400
[10271.585129] R13: 7fd4766252a0 R14: 55f70a0298e0 R15: 07ff
[10271.585221] INFO: task lxcfs:84510 blocked for more than 120 seconds.
[10271.591687]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
[10271.598531] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[10271.606401] lxcfs   D0 84510  1 0x
[10271.606407] Call Trace:
[10271.606416]  __schedule+0x297/0x8b0
[10271.606418]  schedule+0x2c/0x80
[10271.606420]  schedule_preempt_disabled+0xe/0x10
[10271.606421]  __mutex_lock.isra.2+0x18c/0x4d0
[10271.606423]  __mutex_lock_slowpath+0x13/0x20
[10271.606424]  ? __mutex_lock_slowpath+0x13/0x20
[10271.606425]  mutex_lock+0x2f/0x40
[10271.606427]  proc_cgroup_show+0x4c/0x2a0
[10271.606429]  proc_single_show+0x56/0x80
[10271.606432]  seq_read+0xe5/0x430
[10271.606434]  __vfs_read+0x1b/0x40
[10271.606436]  vfs_read+0x8e/0x130
[10271.606437]  SyS_read+0x55/0xc0
[10271.606440]  do_syscall_64+0x73/0x130
[10271.606441]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[10271.606443] RIP: 0033:0x7f1a50e700b4
[10271.606443] RSP: 002b:7f1a2b7fd870 EFLAGS: 0246 ORIG_RAX: 

[10271.606445] RAX: ffda RBX: 0013 RCX: 7f1a50e700b4
[10271.606446] RDX: 0400 RSI: 7f1a04000f90 RDI: 0013
[10271.606446] RBP: 7f1a04000f90 R08: 0001 R09: 
[10271.606447] R10:  R11: 0246 R12: 0400
[10271.606448] R13: 7f1a511482a0 R14:  R15: 7f1a38003a40
[10271.606450] INFO: task lxcfs:84676 blocked for more than 120 seconds.
[10271.612911]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
[10271.619722] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[10271.627577] lxcfs   D0 84676  1 0x
[10271.627579] Call Trace:
[10271.627583]  __schedule+0x297/0x8b0
[10271.627589]  schedule+0x2c/0x80
[10271.627594]  schedule_preempt_disabled+0xe/0x10
[10271.627598]  __mutex_lock.isra.2+0x18c/0x4d0
[10271.627600]  __mutex_lock_slowpath+0x13/0x20
[10271.627601]  ? __mutex_lock_slowpath+0x13/0x20
[10271.627602]  mutex_lock+0x2f/0x40
[10271.627604]  proc_cgroup_show+0x4c/0x2a0
[10271.627606]  proc_single_show+0x56/0x80
[10271.627608]  seq_read+0xe5/0x430
[10271.627610]  __vfs_read+0x1b/0x40
[10271.627611]  vfs_read+0x8e/0x130
[10271.627613]  SyS_read+0x55/0xc0
[10271.627615]  do_syscall_64+0x73/0x130
[10271.627617]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[10271.627618] RIP: 0033:0x7f1a50e700b4
[10271.627618] RSP: 002b:7f1a297f9870 EFLAGS: 0246 ORIG_RAX: 

[10271.627620] RAX: ffda RBX: 0014 RCX: 7f1a50e700b4
[10271.627620] RDX: 0400 RSI: 7f1a2c001e80 RDI: 0014
[10271.627621] RBP: 7f1a2c001e80 R08: 0001 R09: 
[10271.627622] R10:  R11: 0246 R12: 0400
[10271.627622] R13: 7f1a511482a0 R14:  R15: 7f1a2c002b10
[10271.627625] INFO: task lxcfs:115507 blocked for more than 120 seconds.
[10271.634180]   Tainted: P   OE4.15.0-20-generic #21-Ubuntu
[10271.641007] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
message.
[10271.648870] lxcfs   D0 115507  1 0x
[10271.648872] Call Trace:
[10271.648876]  __schedule+0x297/0x8b0
[10271.648883]  schedule+0x2c/0x80

[Kernel-packages] [Bug 1773392] Re: zfs hangs on mount/unmount

2018-07-10 Thread Simos Xenitellis
** Bug watch added: Github Issue Tracker for ZFS #7659
   https://github.com/zfsonlinux/zfs/issues/7659

** Changed in: linux
   Status: Fix Released => Unknown

** Changed in: linux
 Remote watch: Github Issue Tracker for ZFS #7691 => Github Issue Tracker for 
ZFS #7659

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1773392

Title:
  zfs hangs on mount/unmount

Status in Linux:
  Unknown
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  I am running lxd 3.0 on ubuntu 18.04 with kernel 4.15.0-22-generic and
  4.15.0-20-generic (same behaviour) with zfs backend (0.7.5-1ubuntu16;
  also tried 0.7.9).

  Sometimes lxd hangs when I try to stop / restart or "stop && move"
  some containers. Furhter investigation showed that problem is in zfs
  mount or unmount: it just hangs and lxd just wait it. Also commands
  like "zfs list" hangs to.

  It seems that it is not lxd or zfs issue, but kernel bug?
  https://github.com/lxc/lxd/issues/4104#issuecomment-392072939

  I have one test ct that always hangs on restart, so here is info:

  dmesg:
  [ 1330.390938] INFO: task txg_sync:9944 blocked for more than 120 seconds.
  [ 1330.390994]   Tainted: P   O 4.15.0-22-generic #24-Ubuntu
  [ 1330.391044] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [ 1330.391101] txg_syncD0  9944  2 0x8000
  [ 1330.391105] Call Trace:
  [ 1330.391117]  __schedule+0x297/0x8b0
  [ 1330.391122]  schedule+0x2c/0x80
  [ 1330.391136]  cv_wait_common+0x11e/0x140 [spl]
  [ 1330.391141]  ? wait_woken+0x80/0x80
  [ 1330.391152]  __cv_wait+0x15/0x20 [spl]
  [ 1330.391234]  rrw_enter_write+0x3c/0xa0 [zfs]
  [ 1330.391306]  rrw_enter+0x13/0x20 [zfs]
  [ 1330.391380]  spa_sync+0x7c9/0xd80 [zfs]
  [ 1330.391457]  txg_sync_thread+0x2cd/0x4a0 [zfs]
  [ 1330.391534]  ? txg_quiesce_thread+0x3d0/0x3d0 [zfs]
  [ 1330.391543]  thread_generic_wrapper+0x74/0x90 [spl]
  [ 1330.391549]  kthread+0x121/0x140
  [ 1330.391558]  ? __thread_exit+0x20/0x20 [spl]
  [ 1330.391562]  ? kthread_create_worker_on_cpu+0x70/0x70
  [ 1330.391566]  ? kthread_create_worker_on_cpu+0x70/0x70
  [ 1330.391569]  ret_from_fork+0x35/0x40
  [ 1330.391582] INFO: task lxd:12419 blocked for more than 120 seconds.
  [ 1330.391630]   Tainted: P   O 4.15.0-22-generic #24-Ubuntu
  [ 1330.391679] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [ 1330.391735] lxd D0 12419  1 0x
  [ 1330.391739] Call Trace:
  [ 1330.391745]  __schedule+0x297/0x8b0
  [ 1330.391749]  schedule+0x2c/0x80
  [ 1330.391752]  rwsem_down_write_failed+0x162/0x360
  [ 1330.391808]  ? dbuf_rele_and_unlock+0x1a8/0x4b0 [zfs]
  [ 1330.391814]  call_rwsem_down_write_failed+0x17/0x30
  [ 1330.391817]  ? call_rwsem_down_write_failed+0x17/0x30
  [ 1330.391821]  down_write+0x2d/0x40
  [ 1330.391825]  grab_super+0x30/0x90
  [ 1330.391901]  ? zpl_create+0x160/0x160 [zfs]
  [ 1330.391905]  sget_userns+0x91/0x490
  [ 1330.391908]  ? get_anon_bdev+0x100/0x100
  [ 1330.391983]  ? zpl_create+0x160/0x160 [zfs]
  [ 1330.391987]  sget+0x7d/0xa0
  [ 1330.391990]  ? get_anon_bdev+0x100/0x100
  [ 1330.392066]  zpl_mount+0xa8/0x160 [zfs]
  [ 1330.392071]  mount_fs+0x37/0x150
  [ 1330.392077]  vfs_kern_mount.part.23+0x5d/0x110
  [ 1330.392080]  do_mount+0x5ed/0xce0
  [ 1330.392083]  ? copy_mount_options+0x2c/0x220
  [ 1330.392086]  SyS_mount+0x98/0xe0
  [ 1330.392092]  do_syscall_64+0x73/0x130
  [ 1330.392096]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [ 1330.392099] RIP: 0033:0x4db36a
  [ 1330.392101] RSP: 002b:00c4207fa768 EFLAGS: 0216 ORIG_RAX: 
00a5
  [ 1330.392104] RAX: ffda RBX:  RCX: 
004db36a
  [ 1330.392106] RDX: 00c4205984cc RSI: 00c420a6ee00 RDI: 
00c420a23b60
  [ 1330.392108] RBP: 00c4207fa808 R08: 00c4209d4960 R09: 

  [ 1330.392110] R10:  R11: 0216 R12: 

  [ 1330.392112] R13: 0039 R14: 0038 R15: 
0080
  [ 1330.392123] INFO: task lxd:16725 blocked for more than 120 seconds.
  [ 1330.392171]   Tainted: P   O 4.15.0-22-generic #24-Ubuntu
  [ 1330.392220] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [ 1330.392276] lxd D0 16725  1 0x0002
  [ 1330.392279] Call Trace:
  [ 1330.392284]  __schedule+0x297/0x8b0
  [ 1330.392289]  ? irq_work_queue+0x8d/0xa0
  [ 1330.392293]  schedule+0x2c/0x80
  [ 1330.392297]  io_schedule+0x16/0x40
  [ 1330.392302]  wait_on_page_bit_common+0xd8/0x160
  [ 1330.392305]  ? page_cache_tree_insert+0xe0/0xe0
  [ 1330.392309]  __filemap_fdatawait_range+0xfa/0x160
  [ 1330.392313]  ? _cond_resched+0x19/0x40
  [ 1330.392317]  ? bdi_split_work_to_wbs+0x45/0x2c0
  [ 1330.392321]  ? _cond_resched+0x19/0x40
  [ 

[Kernel-packages] [Bug 1607355] Re: Task (usually mongod) blocked more 120 seconds (lock-ups) in juju on lxc/lxd + zfs

2018-07-09 Thread Simos Xenitellis
This might be related to #1773392.

On the other hand, the issue in #1773392 appears to have been introduced
in more recent kernels (16.04 not affected).

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1607355

Title:
  Task (usually mongod) blocked more 120 seconds (lock-ups) in juju on
  lxc/lxd + zfs

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  I was able to reproduce this 2 or 3 times last 2 days. I have the
  following setup:

  Containers for Trusty/kilo service machines:

  inaddy@workstation:~$ lxc-ls  | grep tk
  tkcephmon01  RUNNING 0 -  192.168.65.52 -
  tkcephmon02  RUNNING 0 -  192.168.65.51 -
  tkcephmon03  RUNNING 0 -  192.168.65.48 -
  tkcinder RUNNING 0 -  192.168.65.49 -
  tkdash   RUNNING 0 -  192.168.65.50 -
  tkglance RUNNING 0 -  192.168.65.53 -
  tkjuju   RUNNING 0 -  192.168.65.15 -
  tkkeystone   RUNNING 0 -  192.168.65.54 -
  tkmysql  RUNNING 0 -  192.168.65.55 -
  tknova   RUNNING 0 -  192.168.65.56 -
  tkrabbit RUNNING 0 -  192.168.65.57 -
  tkswiftproxy RUNNING 0 -  192.168.65.58 -

  And compute nodes + neutrongw as kvm guests:

  inaddy@workstation:~$ virsh list --all | grep tk
   21tkcompute01running
   22tkcompute02running
   23tkcompute03running
   24tkneutrongwrunning

  All my LXC containers are on top of ZFS:

  Linux workstation 4.4.0-32-generic #51-Ubuntu SMP Tue Jul 19 18:09:07
  UTC 2016 x86_64 x86_64 x86_64 GNU/Linux

  And my KVM guests are on top of ext4 + 1.2 raid0 stripped volume.

  I'm getting the lockups bellow (usually for mongod, from tkjuju
  container, the juju controller). After the first lockup appears
  (schedule timeout coming from zfs sync logic most likely), JuJu
  controller starts giving me errors on "update-status". From "juju
  status":

  glance/0error  idle1.25.6  10  9292/tcp   
tkglance   hook failed: "update-status"
  keystone/0  error  idle1.25.6  11 
tkkeystone hook failed: "update-status"
  mysql/0 error  idle1.25.6  12 
tkmysqlhook failed: "config-changed"
  neutron-api/0   error  idle1.25.6  4   9696/tcp   
tkneutrongwhook failed: "update-status"
  nova-compute/0  error  idle1.25.6  1  
tkcompute01hook failed: "update-status"
  nova-compute/1  error  idle1.25.6  2  
tkcompute02hook failed: "update-status"
  nova-compute/2  error  idle1.25.6  3  
tkcompute03hook failed: "update-status"

  Lockups:

  [105601.816578] INFO: task mongod:14480 blocked for more than 120 seconds.
  [105601.816583]   Tainted: P   O4.4.0-32-generic #51-Ubuntu
  [105601.816584] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [105601.816586] mongod  D 88010ec47ba8 0 14480  16855 
0x0100
  [105601.816590]  88010ec47ba8 57992eeb 880108e5ee00 
880108e58dc0
  [105601.816592]  88010ec48000 88081ecd6d00 7fff 
8182a600
  [105601.816594]  88010ec47d08 88010ec47bc0 81829e05 

  [105601.816596] Call Trace:
  [105601.816603]  [] ? bit_wait+0x60/0x60
  [105601.816606]  [] schedule+0x35/0x80
  [105601.816608]  [] schedule_timeout+0x1b5/0x270
  [105601.816612]  [] ? find_get_pages_tag+0x109/0x190
  [105601.816614]  [] ? bit_wait+0x60/0x60
  [105601.816616]  [] io_schedule_timeout+0xa4/0x110
  [105601.816618]  [] bit_wait_io+0x1b/0x70
  [105601.816620]  [] __wait_on_bit+0x5d/0x90
  [105601.816622]  [] wait_on_page_bit+0xcb/0xf0
  [105601.816625]  [] ? autoremove_wake_function+0x40/0x40
  [105601.816628]  [] __filemap_fdatawait_range+0xf3/0x160
  [105601.816630]  [] filemap_fdatawait_range+0x14/0x30
  [105601.816632]  [] filemap_write_and_wait_range+0x3f/0x70
  [105601.816682]  [] zpl_fsync+0x38/0x90 [zfs]
  [105601.816685]  [] vfs_fsync_range+0x4b/0xb0
  [105601.816687]  [] SyS_msync+0x17e/0x1f0
  [105601.816689]  [] entry_SYSCALL_64_fastpath+0x16/0x71
  [117121.961545] INFO: task txg_sync:4589 blocked for more than 120 seconds.
  [117121.961549]   Tainted: P   O4.4.0-32-generic #51-Ubuntu
  [117121.961550] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [117121.961551] txg_syncD 8807e1fbbaa8 0  4589  2 
0x
  

[Kernel-packages] [Bug 1773392] Re: zfs hangs on mount/unmount

2018-07-09 Thread Simos Xenitellis
Upstream bug report and pull request to try:

"Kernel error "task zfs:pid blocked for more than 120 seconds" #7691"
https://github.com/zfsonlinux/zfs/issues/7691

"Fix zpl_mount() deadlock #7693"
https://github.com/zfsonlinux/zfs/pull/7693

** Bug watch added: Github Issue Tracker for ZFS #7691
   https://github.com/zfsonlinux/zfs/issues/7691

** Also affects: linux via
   https://github.com/zfsonlinux/zfs/issues/7691
   Importance: Unknown
   Status: Unknown

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1773392

Title:
  zfs hangs on mount/unmount

Status in Linux:
  Unknown
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  I am running lxd 3.0 on ubuntu 18.04 with kernel 4.15.0-22-generic and
  4.15.0-20-generic (same behaviour) with zfs backend (0.7.5-1ubuntu16;
  also tried 0.7.9).

  Sometimes lxd hangs when I try to stop / restart or "stop && move"
  some containers. Furhter investigation showed that problem is in zfs
  mount or unmount: it just hangs and lxd just wait it. Also commands
  like "zfs list" hangs to.

  It seems that it is not lxd or zfs issue, but kernel bug?
  https://github.com/lxc/lxd/issues/4104#issuecomment-392072939

  I have one test ct that always hangs on restart, so here is info:

  dmesg:
  [ 1330.390938] INFO: task txg_sync:9944 blocked for more than 120 seconds.
  [ 1330.390994]   Tainted: P   O 4.15.0-22-generic #24-Ubuntu
  [ 1330.391044] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [ 1330.391101] txg_syncD0  9944  2 0x8000
  [ 1330.391105] Call Trace:
  [ 1330.391117]  __schedule+0x297/0x8b0
  [ 1330.391122]  schedule+0x2c/0x80
  [ 1330.391136]  cv_wait_common+0x11e/0x140 [spl]
  [ 1330.391141]  ? wait_woken+0x80/0x80
  [ 1330.391152]  __cv_wait+0x15/0x20 [spl]
  [ 1330.391234]  rrw_enter_write+0x3c/0xa0 [zfs]
  [ 1330.391306]  rrw_enter+0x13/0x20 [zfs]
  [ 1330.391380]  spa_sync+0x7c9/0xd80 [zfs]
  [ 1330.391457]  txg_sync_thread+0x2cd/0x4a0 [zfs]
  [ 1330.391534]  ? txg_quiesce_thread+0x3d0/0x3d0 [zfs]
  [ 1330.391543]  thread_generic_wrapper+0x74/0x90 [spl]
  [ 1330.391549]  kthread+0x121/0x140
  [ 1330.391558]  ? __thread_exit+0x20/0x20 [spl]
  [ 1330.391562]  ? kthread_create_worker_on_cpu+0x70/0x70
  [ 1330.391566]  ? kthread_create_worker_on_cpu+0x70/0x70
  [ 1330.391569]  ret_from_fork+0x35/0x40
  [ 1330.391582] INFO: task lxd:12419 blocked for more than 120 seconds.
  [ 1330.391630]   Tainted: P   O 4.15.0-22-generic #24-Ubuntu
  [ 1330.391679] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [ 1330.391735] lxd D0 12419  1 0x
  [ 1330.391739] Call Trace:
  [ 1330.391745]  __schedule+0x297/0x8b0
  [ 1330.391749]  schedule+0x2c/0x80
  [ 1330.391752]  rwsem_down_write_failed+0x162/0x360
  [ 1330.391808]  ? dbuf_rele_and_unlock+0x1a8/0x4b0 [zfs]
  [ 1330.391814]  call_rwsem_down_write_failed+0x17/0x30
  [ 1330.391817]  ? call_rwsem_down_write_failed+0x17/0x30
  [ 1330.391821]  down_write+0x2d/0x40
  [ 1330.391825]  grab_super+0x30/0x90
  [ 1330.391901]  ? zpl_create+0x160/0x160 [zfs]
  [ 1330.391905]  sget_userns+0x91/0x490
  [ 1330.391908]  ? get_anon_bdev+0x100/0x100
  [ 1330.391983]  ? zpl_create+0x160/0x160 [zfs]
  [ 1330.391987]  sget+0x7d/0xa0
  [ 1330.391990]  ? get_anon_bdev+0x100/0x100
  [ 1330.392066]  zpl_mount+0xa8/0x160 [zfs]
  [ 1330.392071]  mount_fs+0x37/0x150
  [ 1330.392077]  vfs_kern_mount.part.23+0x5d/0x110
  [ 1330.392080]  do_mount+0x5ed/0xce0
  [ 1330.392083]  ? copy_mount_options+0x2c/0x220
  [ 1330.392086]  SyS_mount+0x98/0xe0
  [ 1330.392092]  do_syscall_64+0x73/0x130
  [ 1330.392096]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [ 1330.392099] RIP: 0033:0x4db36a
  [ 1330.392101] RSP: 002b:00c4207fa768 EFLAGS: 0216 ORIG_RAX: 
00a5
  [ 1330.392104] RAX: ffda RBX:  RCX: 
004db36a
  [ 1330.392106] RDX: 00c4205984cc RSI: 00c420a6ee00 RDI: 
00c420a23b60
  [ 1330.392108] RBP: 00c4207fa808 R08: 00c4209d4960 R09: 

  [ 1330.392110] R10:  R11: 0216 R12: 

  [ 1330.392112] R13: 0039 R14: 0038 R15: 
0080
  [ 1330.392123] INFO: task lxd:16725 blocked for more than 120 seconds.
  [ 1330.392171]   Tainted: P   O 4.15.0-22-generic #24-Ubuntu
  [ 1330.392220] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [ 1330.392276] lxd D0 16725  1 0x0002
  [ 1330.392279] Call Trace:
  [ 1330.392284]  __schedule+0x297/0x8b0
  [ 1330.392289]  ? irq_work_queue+0x8d/0xa0
  [ 1330.392293]  schedule+0x2c/0x80
  [ 1330.392297]  io_schedule+0x16/0x40
  [ 1330.392302]  wait_on_page_bit_common+0xd8/0x160
  [ 1330.392305]  ? page_cache_tree_insert+0xe0/0xe0
  [ 1330.392309]  

[Kernel-packages] [Bug 1611124] Re: W: Possible missing firmware /lib/firmware/i915/kbl_guc_ver9_14.bin for module i915

2017-03-21 Thread Simos Xenitellis
I get this message and I do not even run a Kabylake system ("kbl").

According to https://01.org/linuxgraphics/downloads/firmware

"2. Q: I am seeing the following error message, what do they mean and what 
should I do in response?
   
   W:Possible missing firmware /lib/firmware/i915/kbl_dmc_ver1.bin for module 
i915
 
   A: This warning can be safely ignored."

Therefore, this message can be safely ignored.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-firmware in Ubuntu.
https://bugs.launchpad.net/bugs/1611124

Title:
  W: Possible missing firmware /lib/firmware/i915/kbl_guc_ver9_14.bin
  for module i915

Status in xen:
  Unknown
Status in linux-firmware package in Ubuntu:
  Invalid

Bug description:
  $sudo dpkg -i linux-image-4.8.0-040800rc1-lowlatency

  [...]

  update-initramfs: Generating /boot/initrd.img-4.8.0-040800rc1-lowlatency
  W: Possible missing firmware /lib/firmware/i915/kbl_guc_ver9_14.bin for 
module i915
  W: Possible missing firmware /lib/firmware/i915/bxt_guc_ver8_7.bin for module 
i915

  ProblemType: Bug
  DistroRelease: Ubuntu 16.10
  Package: linux-firmware 1.159
  Uname: Linux 4.8.0-040800rc1-lowlatency x86_64
  ApportVersion: 2.20.3-0ubuntu5
  Architecture: amd64
  Date: Mon Aug  8 17:50:38 2016
  Dependencies:
   
  InstallationDate: Installed on 2015-07-26 (379 days ago)
  InstallationMedia: Ubuntu-GNOME 15.10 "Wily Werewolf" - Alpha amd64 (20150723)
  PackageArchitecture: all
  SourcePackage: linux-firmware
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/xen/+bug/1611124/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 967399] Re: [11.10] Elantech trackpoint does not work Lenovo

2014-12-01 Thread Simos Xenitellis
It's great that it has been fixed upstream.

To backport to 14.04, see instructions at
https://help.ubuntu.com/community/UbuntuBackports

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/967399

Title:
  [11.10] Elantech trackpoint does not work Lenovo

Status in linux package in Ubuntu:
  In Progress
Status in linux source package in Precise:
  Fix Released

Bug description:
  Title: trackpoint does not function

  Summary: trackpoint does not function

  Steps:
  1. log into desktop
  2. use trackpoint and trackpoint buttons

  Expected results: laptop responds to trackpoint and trackpoint buttons
  being pressed

  Actual results: nothing happens
  Manufacturer: LENOVO

  WORKAROUND: sudo modprobe proto=exps enables the trackpoint and
  touchpad, but user will lose multi-touch functionality.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/967399/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1364546] Re: [ASRock Z77 Extreme4] suspend/resume failure [non-free: nvidia]

2014-10-29 Thread Simos Xenitellis
Also occurred on an ASRock Z87 Pro4, after several suspend/resume
cycles. It might be a race condition.

Linux desktop 3.13.0-37-generic #64-Ubuntu SMP Mon Sep 22 21:28:38 UTC
2014 x86_64 x86_64 x86_64 GNU/Linux

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1364546

Title:
  [ASRock Z77 Extreme4] suspend/resume failure [non-free: nvidia]

Status in “linux” package in Ubuntu:
  Incomplete

Bug description:
  Turned off when it should have resumed (keyboard pressed).

  ProblemType: KernelOops
  DistroRelease: Ubuntu 14.10
  Package: linux-image-3.16.0-12-generic 3.16.0-12.17
  ProcVersionSignature: Ubuntu 3.16.0-12.17-generic 3.16.1
  Uname: Linux 3.16.0-12-generic x86_64
  NonfreeKernelModules: nvidia
  Annotation: This occured during a previous suspend and prevented it from 
resuming properly.
  ApportVersion: 2.14.7-0ubuntu1
  Architecture: amd64
  AudioDevicesInUse:
   USERPID ACCESS COMMAND
   /dev/snd/controlC3:  peter  2842 F pulseaudio
   /dev/snd/controlC2:  peter  2842 F pulseaudio
   /dev/snd/controlC0:  peter  2842 F pulseaudio
   /dev/snd/controlC1:  peter  2842 F pulseaudio
  Date: Tue Sep  2 19:55:07 2014
  DuplicateSignature: suspend/resume:To Be Filled By O.E.M. To Be Filled By 
O.E.M.:P2.90
  ExecutablePath: /usr/share/apport/apportcheckresume
  Failure: suspend/resume
  InstallationDate: Installed on 2014-08-16 (16 days ago)
  InstallationMedia: Ubuntu 14.10 Utopic Unicorn - Alpha amd64 (20140816)
  InterpreterPath: /usr/bin/python3.4
  IwConfig:
   eth0  no wireless extensions.
   
   lono wireless extensions.
  MachineType: To Be Filled By O.E.M. To Be Filled By O.E.M.
  ProcCmdline: /usr/bin/python3 /usr/share/apport/apportcheckresume
  ProcEnviron:
   TERM=linux
   PATH=(custom, no user)
  ProcFB: 0 inteldrmfb
  ProcKernelCmdLine: BOOT_IMAGE=/boot/vmlinuz-3.16.0-12-generic 
root=UUID=a01a8a8f-61c7-4592-a658-643a4dcf297d ro quiet splash
  PulseList: Error: command ['pacmd', 'list'] failed with exit code 1: No 
PulseAudio daemon running, or not running as session daemon.
  RelatedPackageVersions:
   linux-restricted-modules-3.16.0-12-generic N/A
   linux-backports-modules-3.16.0-12-generic  N/A
   linux-firmware 1.133
  RfKill:
   
  SourcePackage: linux
  Title: [To Be Filled By O.E.M. To Be Filled By O.E.M.] suspend/resume failure 
[non-free: nvidia]
  UpgradeStatus: No upgrade log present (probably fresh install)
  UserGroups:
   
  WifiSyslog:
   
  dmi.bios.date: 07/11/2013
  dmi.bios.vendor: American Megatrends Inc.
  dmi.bios.version: P2.90
  dmi.board.name: Z77 Extreme4
  dmi.board.vendor: ASRock
  dmi.chassis.asset.tag: To Be Filled By O.E.M.
  dmi.chassis.type: 3
  dmi.chassis.vendor: To Be Filled By O.E.M.
  dmi.chassis.version: To Be Filled By O.E.M.
  dmi.modalias: 
dmi:bvnAmericanMegatrendsInc.:bvrP2.90:bd07/11/2013:svnToBeFilledByO.E.M.:pnToBeFilledByO.E.M.:pvrToBeFilledByO.E.M.:rvnASRock:rnZ77Extreme4:rvr:cvnToBeFilledByO.E.M.:ct3:cvrToBeFilledByO.E.M.:
  dmi.product.name: To Be Filled By O.E.M.
  dmi.product.version: To Be Filled By O.E.M.
  dmi.sys.vendor: To Be Filled By O.E.M.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1364546/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1364546] Re: [ASRock Z77 Extreme4] suspend/resume failure [non-free: nvidia]

2014-10-29 Thread Simos Xenitellis
Christopher, the motherboard is from a LoCo member and I do not have access to 
it at the moment.
What is common between the two cases, is the motherboard being ASRock.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1364546

Title:
  [ASRock Z77 Extreme4] suspend/resume failure [non-free: nvidia]

Status in “linux” package in Ubuntu:
  Incomplete

Bug description:
  Turned off when it should have resumed (keyboard pressed).

  ProblemType: KernelOops
  DistroRelease: Ubuntu 14.10
  Package: linux-image-3.16.0-12-generic 3.16.0-12.17
  ProcVersionSignature: Ubuntu 3.16.0-12.17-generic 3.16.1
  Uname: Linux 3.16.0-12-generic x86_64
  NonfreeKernelModules: nvidia
  Annotation: This occured during a previous suspend and prevented it from 
resuming properly.
  ApportVersion: 2.14.7-0ubuntu1
  Architecture: amd64
  AudioDevicesInUse:
   USERPID ACCESS COMMAND
   /dev/snd/controlC3:  peter  2842 F pulseaudio
   /dev/snd/controlC2:  peter  2842 F pulseaudio
   /dev/snd/controlC0:  peter  2842 F pulseaudio
   /dev/snd/controlC1:  peter  2842 F pulseaudio
  Date: Tue Sep  2 19:55:07 2014
  DuplicateSignature: suspend/resume:To Be Filled By O.E.M. To Be Filled By 
O.E.M.:P2.90
  ExecutablePath: /usr/share/apport/apportcheckresume
  Failure: suspend/resume
  InstallationDate: Installed on 2014-08-16 (16 days ago)
  InstallationMedia: Ubuntu 14.10 Utopic Unicorn - Alpha amd64 (20140816)
  InterpreterPath: /usr/bin/python3.4
  IwConfig:
   eth0  no wireless extensions.
   
   lono wireless extensions.
  MachineType: To Be Filled By O.E.M. To Be Filled By O.E.M.
  ProcCmdline: /usr/bin/python3 /usr/share/apport/apportcheckresume
  ProcEnviron:
   TERM=linux
   PATH=(custom, no user)
  ProcFB: 0 inteldrmfb
  ProcKernelCmdLine: BOOT_IMAGE=/boot/vmlinuz-3.16.0-12-generic 
root=UUID=a01a8a8f-61c7-4592-a658-643a4dcf297d ro quiet splash
  PulseList: Error: command ['pacmd', 'list'] failed with exit code 1: No 
PulseAudio daemon running, or not running as session daemon.
  RelatedPackageVersions:
   linux-restricted-modules-3.16.0-12-generic N/A
   linux-backports-modules-3.16.0-12-generic  N/A
   linux-firmware 1.133
  RfKill:
   
  SourcePackage: linux
  Title: [To Be Filled By O.E.M. To Be Filled By O.E.M.] suspend/resume failure 
[non-free: nvidia]
  UpgradeStatus: No upgrade log present (probably fresh install)
  UserGroups:
   
  WifiSyslog:
   
  dmi.bios.date: 07/11/2013
  dmi.bios.vendor: American Megatrends Inc.
  dmi.bios.version: P2.90
  dmi.board.name: Z77 Extreme4
  dmi.board.vendor: ASRock
  dmi.chassis.asset.tag: To Be Filled By O.E.M.
  dmi.chassis.type: 3
  dmi.chassis.vendor: To Be Filled By O.E.M.
  dmi.chassis.version: To Be Filled By O.E.M.
  dmi.modalias: 
dmi:bvnAmericanMegatrendsInc.:bvrP2.90:bd07/11/2013:svnToBeFilledByO.E.M.:pnToBeFilledByO.E.M.:pvrToBeFilledByO.E.M.:rvnASRock:rnZ77Extreme4:rvr:cvnToBeFilledByO.E.M.:ct3:cvrToBeFilledByO.E.M.:
  dmi.product.name: To Be Filled By O.E.M.
  dmi.product.version: To Be Filled By O.E.M.
  dmi.sys.vendor: To Be Filled By O.E.M.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1364546/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1271268] Re: 174c:5106 1 TB External USB 3.0 Drive Fails to Automount through USB 3.0 dock with XHCI Enabled

2014-06-19 Thread Simos Xenitellis
Apparently this is a regression in the xhci code (USB3). If you connect
the device to a USB2 port (if you have one) or disable xHCI in the BIOS,
the problem is gone. Of course you get USB2 speeds as well.

Relevant bug reports:

Bug#741989: linux: 3.13.5: known regression: xHCI xhci_drop_endpoint called 
with disabled ep
https://lists.debian.org/debian-kernel/2014/03/msg00153.html

Bug 864866 - USB3 device randomly disconnecting while in use or not
https://bugzilla.novell.com/show_bug.cgi?id=864866

USB 3.0 corrupts NTFS disks
https://forums.gentoo.org/viewtopic-t-963046-start-0.html

In a nutshell, 
1. when something goes bad, the first dmesg messages say

[ 3888.906845] usb 4-4: reset SuperSpeed USB device number 3 using xhci_hcd
[ 3888.923000] xhci_hcd :00:14.0: xHCI xhci_drop_endpoint called with 
disabled ep 8801de04f680
[ 3888.923002] xhci_hcd :00:14.0: xHCI xhci_drop_endpoint called with 
disabled ep 8801de04f6c0
(probably multiple times if keep using the drive)

2. Then, the drive is disconnected and reconnected again. If you have
the default setting to automount, you get many Nautilus windows opening
up.

[ 4676.122427] usb 4-4: USB disconnect, device number 3
[ 4676.490858] sd 6:0:0:0: [sdb] Synchronizing SCSI cache
[ 4676.490878] sd 6:0:0:0: [sdb]  
[ 4676.490879] Result: hostbyte=DID_NO_CONNECT driverbyte=DRIVER_OK
[ 4676.490934] usb 4-4: Set SEL for device-initiated U1 failed.
[ 4676.490935] usb 4-4: Set SEL for device-initiated U2 failed.

See http://blogs.msdn.com/b/usbcoreblog/archive/2012/07/13/common-
issues-in-usb-3-0-devices.aspx for more about Set SEL.

3. After some more use, the device gives buffer I/O errors,

[ 5398.216920] Buffer I/O error on device sdb1, logical block 91570828
[ 5398.216933] Buffer I/O error on device sdb1, logical block 91570828
[ 5398.20] Buffer I/O error on device sdb1, logical block 91570828


** Bug watch added: Novell/SUSE Bugzilla #864866
   https://bugzilla.novell.com/show_bug.cgi?id=864866

** Also affects: linux (Suse) via
   https://bugzilla.novell.com/show_bug.cgi?id=864866
   Importance: Unknown
   Status: Unknown

** Bug watch added: Debian Bug tracker #741989
   http://bugs.debian.org/cgi-bin/bugreport.cgi?bug=741989

** Also affects: linux (Debian) via
   http://bugs.debian.org/cgi-bin/bugreport.cgi?bug=741989
   Importance: Unknown
   Status: Unknown

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1271268

Title:
  174c:5106 1 TB External USB 3.0 Drive Fails to Automount through USB
  3.0 dock with XHCI Enabled

Status in “linux” package in Ubuntu:
  Triaged
Status in “linux” package in Debian:
  Unknown
Status in “linux” package in Suse:
  Unknown

Bug description:
  I have a 1 TB Western Digital drive in a USB 3.0 HDD dock that will not 
automount with Ubuntu 13.10 while XHCI is enabled in the BIOS. If I disable 
XHCI in the BIOS, it automounts normally. I've tried an external USB 3.0 
enclosure and two different brands of USB 3.0 HDD docks (that use different 
chipsets) with the 1 TB drive and it still will not automount. Dmesg shows the 
following when the drive is turned on:
  [10813.786820] usb 4-5.3: new SuperSpeed USB device number 4 using xhci_hcd
  [10823.816236] usb 4-5.3: New USB device found, idVendor=174c, idProduct=5106
  [10823.816245] usb 4-5.3: New USB device strings: Mfr=2, Product=3, 
SerialNumber=1
  [10823.816250] usb 4-5.3: Product: AS2105
  [10823.816255] usb 4-5.3: Manufacturer: ASMedia
  [10828.823853] usb 4-5.3: can't set config #1, error -110

  WORKAROUND: Disable XHCI in the BIOS, but this runs at USB 2.0 speeds
  using EHCI.

  I actually thought the 1 TB drive was bad, and I RMA'd it to Western
  Digital, but the replacement drive fails with the same error message.
  I have a couple of Western Digital 500 GB drives that work just fine
  in the same HDD dock with XHCI enabled. All the drives were formatted
  in Ubuntu 13.10 using LUKS via the Disks Utility. I have tried various
  settings in the BIOS (XHCI set at Smart Auto, Auto, and Manual) but
  they make no difference. I found an old bug report with the same
  can't set config #1, error -110 error message that recommended
  adding pci=nomsi to Grub, but that did not work either.

  Upstream URL: permalink.gmane.org/gmane.linux.usb.general/101979

  ProblemType: Bug
  DistroRelease: Ubuntu 13.10
  Package: linux-image-3.11.0-15-generic 3.11.0-15.23
  ProcVersionSignature: Ubuntu 3.11.0-15.23-generic 3.11.10
  Uname: Linux 3.11.0-15-generic x86_64
  ApportVersion: 2.12.5-0ubuntu2.2
  Architecture: amd64
  AudioDevicesInUse:
   USERPID ACCESS COMMAND
   /dev/snd/controlC1:  jsherman   2017 F pulseaudio
   /dev/snd/controlC0:  jsherman   2017 F pulseaudio
   /dev/snd/pcmC0D1p:   jsherman   2017 F...m pulseaudio
  CurrentDmesg:
   [   81.782320] e1000e :00:19.0: irq 44 for MSI/MSI-X
   [   81.782396] IPv6: