[Kernel-packages] [Bug 1801349] Re: zpool create -f lxd /dev/vdb fails on cosmic (18.10) -- func27

2018-11-07 Thread Colin Ian King
Can you supply me the commands required to reproduce this issue, I'm not
familiar with the test environment you are using.

Thanks

** Changed in: charm-lxd
   Status: Incomplete => In Progress

** Changed in: charm-lxd
   Status: In Progress => New

** Changed in: zfs-linux (Ubuntu)
   Status: New => In Progress

** Changed in: zfs-linux (Ubuntu)
   Importance: Undecided => High

** Changed in: zfs-linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: charm-lxd
 Assignee: Colin Ian King (colin-king) => (unassigned)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1801349

Title:
  zpool create -f lxd /dev/vdb fails on cosmic (18.10) -- func27

Status in OpenStack LXD Charm:
  New
Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  Test: tests/gate-basic-cosmic-rocky

  As part of the config, the lxd charm creates a pool device depending
  on the config.  The test config is:

  lxd_config = {
  'block-devices': '/dev/vdb',
  'ephemeral-unmount': '/mnt',
  'storage-type': 'zfs',
  'overwrite': True
  }

  The config drive is normally mounted on /mnt, and the lxd charm
  umounts it as part of the start up.  The /etc/fstab on the unit is:

  # cat /etc/fstab 
  LABEL=cloudimg-rootfs   /ext4   defaults0 0
  LABEL=UEFI  /boot/efi   vfatdefaults0 0
  /dev/vdb/mntauto
defaults,nofail,x-systemd.requires=cloud-init.service,comment=cloudconfig   
0   2
  /dev/vdcnoneswapsw,comment=cloudconfig  0   0

  
  However, even after umount-ing the /mnt off of /dev/vdb, the zpool create 
command still fails:

  # zpool create -f lxd /dev/vdb
  /dev/vdb is in use and contains a unknown filesystem.

  
  If the /etc/fstab is edited so that /dev/vdb is *never* mounted and then 
rebooted, then the zpool create command succeeds:

  # zpool list
  NAME   SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  lxd   14.9G   106K  14.9G - 0% 0%  1.00x  ONLINE  -

  # zpool status lxd
pool: lxd
   state: ONLINE
scan: none requested
  config:

  NAMESTATE READ WRITE CKSUM
  lxd ONLINE   0 0 0
vdb   ONLINE   0 0 0

  errors: No known data errors

  Something odd is going on with cosmic (18.10) and the combination of
  lxd, zfs and the kernel

  lxd version: 3.6
  zfsutils-linux/cosmic,now 0.7.9-3ubuntu6
  Linux: 4.18.0-10-generic

To manage notifications about this bug go to:
https://bugs.launchpad.net/charm-lxd/+bug/1801349/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1801856] Re: seccomp in ubuntu_stress_smoke_test failed on ARM64/Power8 Trusty

2018-11-07 Thread Colin Ian King
The arches don't have HAVE_ARCH_SECCOMP_FILTER enabled as this has not
been implemented, so seccomp filtering does not exist. I'll fix stress-
ng to handle this corner case without an error.

** Changed in: stress-ng
   Status: New => In Progress

** Changed in: stress-ng
   Importance: Undecided => High

** Changed in: stress-ng
 Assignee: (unassigned) => Colin Ian King (colin-king)

** No longer affects: linux (Ubuntu)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1801856

Title:
  seccomp in ubuntu_stress_smoke_test failed on ARM64/Power8 Trusty

Status in Stress-ng:
  In Progress
Status in ubuntu-kernel-tests:
  Confirmed

Bug description:
  Didn't see this on amd64 / i386, but it can be reproduced on power8
  and ARM64

seccomp STARTING
seccomp RETURNED 2
seccomp FAILED
stress-ng: debug: [7528] 8 processors online, 8 processors configured
stress-ng: info:  [7528] dispatching hogs: 4 seccomp
stress-ng: debug: [7528] /sys/devices/system/cpu/cpu0/cache does not exist
stress-ng: info:  [7528] cache allocate: using built-in defaults as unable 
to determine cache details
stress-ng: debug: [7528] cache allocate: default cache size: 2048K
stress-ng: debug: [7528] starting stressors
stress-ng: debug: [7529] stress-ng-seccomp: started [7529] (instance 0)
stress-ng: debug: [7528] 4 stressors spawned
stress-ng: debug: [7530] stress-ng-seccomp: started [7530] (instance 1)
stress-ng: fail:  [7533] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: fail:  [7534] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: debug: [7531] stress-ng-seccomp: started [7531] (instance 2)
stress-ng: fail:  [7530] stress-ng-seccomp: aborting because of unexpected 
failure in child process
stress-ng: debug: [7530] stress-ng-seccomp: exited [7530] (instance 1)
stress-ng: fail:  [7529] stress-ng-seccomp: aborting because of unexpected 
failure in child process
stress-ng: debug: [7529] stress-ng-seccomp: exited [7529] (instance 0)
stress-ng: debug: [7532] stress-ng-seccomp: started [7532] (instance 3)
stress-ng: error: [7528] process 7529 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7529] terminated
stress-ng: error: [7528] process 7530 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7530] terminated
stress-ng: fail:  [7535] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: fail:  [7536] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: fail:  [7532] stress-ng-seccomp: aborting because of unexpected 
failure in child process
stress-ng: debug: [7531] stress-ng-seccomp: exited [7531] (instance 2)
stress-ng: error: [7528] process 7531 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7531] terminated
stress-ng: error: [7528] process 7532 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7532] terminated
stress-ng: info:  [7528] unsuccessful run completed in 0.00s

  ProblemType: Bug
  DistroRelease: Ubuntu 14.04
  Package: linux-image-3.13.0-161-generic 3.13.0-161.211
  ProcVersionSignature: User Name 3.13.0-161.211-generic 3.13.11-ckt39
  Uname: Linux 3.13.0-161-generic aarch64
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access /dev/snd/: No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.14.1-0ubuntu3.29
  Architecture: arm64
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: [Errno 2] No such file or directory: 'iw'
  CurrentDmesg:
   
  Date: Tue Nov  6 06:23:27 2018
  IwConfig:
   lono wireless extensions.
   
   enp1s0no wireless extensions.
   
   enp1s0d1  no wireless extensions.
  Lsusb: Error: command ['lsusb'] failed with exit code 1: unable to initialize 
libusb: -99
  PciMultimedia:
   
  ProcFB:
   
  ProcKernelCmdLine: console=ttyS0,9600n8r ro
  RelatedPackageVersions:
   linux-restricted-modules-3.13.0-161-generic N/A
   linux-backports-modules-3.13.0-161-generic  N/A
   linux-firmware  1.127.24
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/stress-ng/+bug/1801856/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.laun

[Kernel-packages] [Bug 1801856] Re: seccomp in ubuntu_stress_smoke_test failed on ARM64/Power8 Trusty

2018-11-07 Thread Colin Ian King
CONFIG_SECCOMP_FILTER does not appear to be configured for the ARM64
system, I presume this is the same for the Power8 box too.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1801856

Title:
  seccomp in ubuntu_stress_smoke_test failed on ARM64/Power8 Trusty

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  Didn't see this on amd64 / i386, but it can be reproduced on power8
  and ARM64

seccomp STARTING
seccomp RETURNED 2
seccomp FAILED
stress-ng: debug: [7528] 8 processors online, 8 processors configured
stress-ng: info:  [7528] dispatching hogs: 4 seccomp
stress-ng: debug: [7528] /sys/devices/system/cpu/cpu0/cache does not exist
stress-ng: info:  [7528] cache allocate: using built-in defaults as unable 
to determine cache details
stress-ng: debug: [7528] cache allocate: default cache size: 2048K
stress-ng: debug: [7528] starting stressors
stress-ng: debug: [7529] stress-ng-seccomp: started [7529] (instance 0)
stress-ng: debug: [7528] 4 stressors spawned
stress-ng: debug: [7530] stress-ng-seccomp: started [7530] (instance 1)
stress-ng: fail:  [7533] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: fail:  [7534] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: debug: [7531] stress-ng-seccomp: started [7531] (instance 2)
stress-ng: fail:  [7530] stress-ng-seccomp: aborting because of unexpected 
failure in child process
stress-ng: debug: [7530] stress-ng-seccomp: exited [7530] (instance 1)
stress-ng: fail:  [7529] stress-ng-seccomp: aborting because of unexpected 
failure in child process
stress-ng: debug: [7529] stress-ng-seccomp: exited [7529] (instance 0)
stress-ng: debug: [7532] stress-ng-seccomp: started [7532] (instance 3)
stress-ng: error: [7528] process 7529 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7529] terminated
stress-ng: error: [7528] process 7530 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7530] terminated
stress-ng: fail:  [7535] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: fail:  [7536] stress-ng-seccomp: prctl PR_SET_SECCOMP failed, 
errno=22 (Invalid argument)
stress-ng: fail:  [7532] stress-ng-seccomp: aborting because of unexpected 
failure in child process
stress-ng: debug: [7531] stress-ng-seccomp: exited [7531] (instance 2)
stress-ng: error: [7528] process 7531 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7531] terminated
stress-ng: error: [7528] process 7532 (stress-ng-seccomp) terminated with 
an error, exit status=1 (stress-ng core failure)
stress-ng: debug: [7528] process [7532] terminated
stress-ng: info:  [7528] unsuccessful run completed in 0.00s

  ProblemType: Bug
  DistroRelease: Ubuntu 14.04
  Package: linux-image-3.13.0-161-generic 3.13.0-161.211
  ProcVersionSignature: User Name 3.13.0-161.211-generic 3.13.11-ckt39
  Uname: Linux 3.13.0-161-generic aarch64
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access /dev/snd/: No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.14.1-0ubuntu3.29
  Architecture: arm64
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: [Errno 2] No such file or directory: 'iw'
  CurrentDmesg:
   
  Date: Tue Nov  6 06:23:27 2018
  IwConfig:
   lono wireless extensions.
   
   enp1s0no wireless extensions.
   
   enp1s0d1  no wireless extensions.
  Lsusb: Error: command ['lsusb'] failed with exit code 1: unable to initialize 
libusb: -99
  PciMultimedia:
   
  ProcFB:
   
  ProcKernelCmdLine: console=ttyS0,9600n8r ro
  RelatedPackageVersions:
   linux-restricted-modules-3.13.0-161-generic N/A
   linux-backports-modules-3.13.0-161-generic  N/A
   linux-firmware  1.127.24
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/stress-ng/+bug/1801856/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1801865] Re: af-alg in ubuntu_stress_smoke_test failed with Trusty Power8

2018-11-06 Thread Colin Ian King
ENOMEM is not a critical failure, adding a skip on the test if sendmsg
fails.

Commit pushed:

http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=55375dcae101a995dc1d932acd90b86f1ea7b4a7

** Changed in: stress-ng
   Importance: Undecided => High

** Changed in: stress-ng
   Status: New => In Progress

** Changed in: stress-ng
 Assignee: (unassigned) => Colin Ian King (colin-king)

** No longer affects: linux (Ubuntu)

** Changed in: stress-ng
   Status: In Progress => Fix Committed

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1801865

Title:
  af-alg in ubuntu_stress_smoke_test failed with Trusty Power8

Status in Stress-ng:
  Fix Committed
Status in ubuntu-kernel-tests:
  New

Bug description:
  The af-alg test failed with Trusty 3.13.0-161-generic Power8,
  reproduce rate 5/5

  06:48:39 DEBUG| [stdout] af-alg STARTING
  06:48:39 DEBUG| [stdout] af-alg RETURNED 2
  06:48:39 DEBUG| [stdout] af-alg FAILED
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] 128 processors online, 128 
processors configured
  06:48:39 DEBUG| [stdout] stress-ng: info:  [19624] dispatching hogs: 4 af-alg
  06:48:39 DEBUG| [stdout] stress-ng: info:  [19624] cache allocate: using 
built-in defaults as unable to determine cache details
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] cache allocate: default 
cache size: 2048K
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] starting stressors
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] 4 stressors spawned
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19625] stress-ng-af-alg: started 
[19625] (instance 0)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19626] stress-ng-af-alg: started 
[19626] (instance 1)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19627] stress-ng-af-alg: started 
[19627] (instance 2)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19628] stress-ng-af-alg: started 
[19628] (instance 3)
  06:48:39 DEBUG| [stdout] stress-ng: fail:  [19628] stress-ng-af-alg: sendmsg 
using cbc(aes) failed: errno=12 (Cannot allocate memory)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19626] stress-ng-af-alg: exited 
[19626] (instance 1)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19628] stress-ng-af-alg: exited 
[19628] (instance 3)
  06:48:39 DEBUG| [stdout] stress-ng: fail:  [19627] stress-ng-af-alg: sendmsg 
using cbc(aes) failed: errno=12 (Cannot allocate memory)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19625] stress-ng-af-alg: exited 
[19625] (instance 0)
  06:48:39 DEBUG| [stdout] stress-ng: error: [19624] process 19625 
(stress-ng-af-alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] process [19625] terminated
  06:48:39 DEBUG| [stdout] stress-ng: error: [19624] process 19626 
(stress-ng-af-alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] process [19626] terminated
  06:48:39 DEBUG| [stdout] stress-ng: error: [19624] process 19627 
(stress-ng-af-alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] process [19627] terminated
  06:48:39 DEBUG| [stdout] stress-ng: error: [19624] process 19628 
(stress-ng-af-alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  06:48:39 DEBUG| [stdout] stress-ng: debug: [19624] process [19628] terminated
  06:48:39 DEBUG| [stdout] stress-ng: info:  [19624] unsuccessful run completed 
in 0.08s
  06:48:39 DEBUG| [stdout]

To manage notifications about this bug go to:
https://bugs.launchpad.net/stress-ng/+bug/1801865/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1801347] Re: ubuntu_vfat_stress will pass even with error reported by stress-ng

2018-11-02 Thread Colin Ian King
Fix committed: http://kernel.ubuntu.com/git/ubuntu/autotest-client-
tests.git/commit/?id=7358ec3b7ffc243168f22da482377c16902320f0

** No longer affects: stress-ng

** Changed in: ubuntu-kernel-tests
   Status: Invalid => Fix Committed

** Changed in: ubuntu-kernel-tests
   Importance: Undecided => Medium

** Changed in: ubuntu-kernel-tests
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: ubuntu-kernel-tests
   Importance: Medium => High

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1801347

Title:
  ubuntu_vfat_stress will pass even with error reported by stress-ng

Status in ubuntu-kernel-tests:
  Fix Committed
Status in linux package in Ubuntu:
  Invalid

Bug description:
  The test will pass even with stress-ng complaining about "No space
  left on device", not sure if this is an expected behaviour, but it
  might be confusing to testers.

   stress-ng: info:  [36588] dispatching hogs: 2 hdd, 2 lockf, 2 seek, 2 aio, 2 
dentry, 2 dir, 2 fallocate, 2 fstat, 2 lease, 2 open, 2 rename, 2 chdir, 2 
rename
   stress-ng: fail:  [36609] stress-ng-chdir: mkdir failed, errno=28 (No space 
left on device)
   stress-ng: fail:  [36623] stress-ng-chdir: mkdir failed, errno=28 (No space 
left on device)
   stress-ng: fail:  [36611] stress-ng-hdd: read failed, errno=28 (No space 
left on device)
   stress-ng: error: [36588] process 36611 (stress-ng-hdd) terminated with an 
error, exit status=1 (stress-ng core failure)
   stress-ng: error: [36588] process [36609] (stress-ng-chdir) aborted early, 
out of system resources
   stress-ng: error: [36588] process [36623] (stress-ng-chdir) aborted early, 
out of system resources
   stress-ng: info:  [36588] unsuccessful run completed in 14.35s
   stress-ng: info:  [36588] stressor   bogo ops real time  usr time  sys 
time   bogo ops/s   bogo ops/s
   stress-ng: info:  [36588]   (secs)(secs)
(secs)   (real time) (usr+sys time)
   stress-ng: info:  [36588] hdd   16391  7.39  0.44  
6.44  2218.21  2382.41
   stress-ng: info:  [36588] lockf131604 10.16  0.05 
12.43 12947.00 10545.19
   stress-ng: info:  [36588] seek1987284 10.17  1.33 
10.40195448.06169418.93
   stress-ng: info:  [36588] aio2496 10.15  0.00  
0.00   245.83 0.00
   stress-ng: info:  [36588] dentry  171 10.42  0.00  
0.6116.41   280.33
   stress-ng: info:  [36588] dir2539 14.09  0.00 
15.40   180.14   164.87
   stress-ng: info:  [36588] fallocate   170 10.17  0.04  
5.3516.7131.54
   stress-ng: info:  [36588] fstat  2147 10.00  0.02  
0.26   214.64  7667.86
   stress-ng: info:  [36588] lease66 10.14  0.61  
5.15 6.5111.46
   stress-ng: info:  [36588] open1434208 10.11  0.44 
12.20141807.28113465.82
   stress-ng: info:  [36588] rename 6526 10.01  0.00  
2.94   651.86  2219.73
   stress-ng: info:  [36588] rename 6820 10.00  0.00  
2.49   681.85  2738.96
   stress-ng: info:  [36588] for a 14.35s run time:
   stress-ng: info:  [36588] 114.80s available CPU time
   stress-ng: info:  [36588]   3.05s user time   (  2.66%)
   stress-ng: info:  [36588]  75.50s system time ( 65.77%)
   stress-ng: info:  [36588]  78.55s total time  ( 68.43%)
   stress-ng: info:  [36588] load average: 17.49 10.98 4.90
   Stress-ng stressor failed, error: 2
   umounting vfat /mnt/vfat-test-29590
   umounting tmpfs /mnt/vfat-test-29590
   


   Completed

   Kernel issues: NONE

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-139-generic 4.4.0-139.165
  ProcVersionSignature: Ubuntu 4.4.0-139.165-generic 4.4.160
  Uname: Linux 4.4.0-139-generic s390x
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  CurrentDmesg:
   
  Date: Fri Nov  2 08:20:40 2018
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig'
  Lspci:
   
  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:
   
  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKe

[Kernel-packages] [Bug 1801347] Re: ubuntu_vfat_stress will pass even with error reported by stress-ng

2018-11-02 Thread Colin Ian King
This is intentional behavior as the test is seeing how to deal with file
operations when a block device is full.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1801347

Title:
  ubuntu_vfat_stress will pass even with error reported by stress-ng

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  The test will pass even with stress-ng complaining about "No space
  left on device", not sure if this is an expected behaviour, but it
  might be confusing to testers.

   stress-ng: info:  [36588] dispatching hogs: 2 hdd, 2 lockf, 2 seek, 2 aio, 2 
dentry, 2 dir, 2 fallocate, 2 fstat, 2 lease, 2 open, 2 rename, 2 chdir, 2 
rename
   stress-ng: fail:  [36609] stress-ng-chdir: mkdir failed, errno=28 (No space 
left on device)
   stress-ng: fail:  [36623] stress-ng-chdir: mkdir failed, errno=28 (No space 
left on device)
   stress-ng: fail:  [36611] stress-ng-hdd: read failed, errno=28 (No space 
left on device)
   stress-ng: error: [36588] process 36611 (stress-ng-hdd) terminated with an 
error, exit status=1 (stress-ng core failure)
   stress-ng: error: [36588] process [36609] (stress-ng-chdir) aborted early, 
out of system resources
   stress-ng: error: [36588] process [36623] (stress-ng-chdir) aborted early, 
out of system resources
   stress-ng: info:  [36588] unsuccessful run completed in 14.35s
   stress-ng: info:  [36588] stressor   bogo ops real time  usr time  sys 
time   bogo ops/s   bogo ops/s
   stress-ng: info:  [36588]   (secs)(secs)
(secs)   (real time) (usr+sys time)
   stress-ng: info:  [36588] hdd   16391  7.39  0.44  
6.44  2218.21  2382.41
   stress-ng: info:  [36588] lockf131604 10.16  0.05 
12.43 12947.00 10545.19
   stress-ng: info:  [36588] seek1987284 10.17  1.33 
10.40195448.06169418.93
   stress-ng: info:  [36588] aio2496 10.15  0.00  
0.00   245.83 0.00
   stress-ng: info:  [36588] dentry  171 10.42  0.00  
0.6116.41   280.33
   stress-ng: info:  [36588] dir2539 14.09  0.00 
15.40   180.14   164.87
   stress-ng: info:  [36588] fallocate   170 10.17  0.04  
5.3516.7131.54
   stress-ng: info:  [36588] fstat  2147 10.00  0.02  
0.26   214.64  7667.86
   stress-ng: info:  [36588] lease66 10.14  0.61  
5.15 6.5111.46
   stress-ng: info:  [36588] open1434208 10.11  0.44 
12.20141807.28113465.82
   stress-ng: info:  [36588] rename 6526 10.01  0.00  
2.94   651.86  2219.73
   stress-ng: info:  [36588] rename 6820 10.00  0.00  
2.49   681.85  2738.96
   stress-ng: info:  [36588] for a 14.35s run time:
   stress-ng: info:  [36588] 114.80s available CPU time
   stress-ng: info:  [36588]   3.05s user time   (  2.66%)
   stress-ng: info:  [36588]  75.50s system time ( 65.77%)
   stress-ng: info:  [36588]  78.55s total time  ( 68.43%)
   stress-ng: info:  [36588] load average: 17.49 10.98 4.90
   Stress-ng stressor failed, error: 2
   umounting vfat /mnt/vfat-test-29590
   umounting tmpfs /mnt/vfat-test-29590
   


   Completed

   Kernel issues: NONE

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-139-generic 4.4.0-139.165
  ProcVersionSignature: Ubuntu 4.4.0-139.165-generic 4.4.160
  Uname: Linux 4.4.0-139-generic s390x
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  CurrentDmesg:
   
  Date: Fri Nov  2 08:20:40 2018
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig'
  Lspci:
   
  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:
   
  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=UUID=44b0b919-a1a4-4849-9425-e71d4ac87d85 
crashkernel=196M nobp=1 BOOT_IMAGE=0
  RelatedPackageVersions:
   linux-restricted-modules-4.4.0-139-generic N/A
   linux-backports-modules-4.4.0-139-generic  N/A
   linux-firmware 1.157.20
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  

[Kernel-packages] [Bug 1800609] Re: dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on LPAR

2018-10-30 Thread Colin Ian King
** No longer affects: linux (Ubuntu)

** Summary changed:

- dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on LPAR
+ dev test in ubuntu_stress_smoke_tests throws usercopy warning s390x Ubuntu on 
LPAR

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1800609

Title:
  dev test in ubuntu_stress_smoke_tests throws usercopy warning s390x
  Ubuntu on LPAR

Status in Stress-ng:
  Fix Committed
Status in ubuntu-kernel-tests:
  New

Bug description:
  The dev test on a s390x node (Ubuntu on LPAR) will hang, kernel bug
  could be found in the syslog

  [54084.082277] [ cut here ]
  [54084.082279] kernel BUG at mm/usercopy.c:100!
  [54084.082294] illegal operation: 0001 ilc:1 [#256] SMP 
  [54084.082297] Modules linked in: vhost_net vhost tap cuse dccp_ipv4 dccp 
algif_rng salsa20_generic camellia_generic cast6_generic cast_common 
serpent_generic twofish_generic twofish_common lrw algif_skcipher tgr192 wp512 
rmd320 rmd256 rmd160 rmd128 md4 algif_hash af_alg aufs ip6t_MASQUERADE 
ip6table_nat ip6table_filter ip6_tables xt_comment xt_CHECKSUM iptable_mangle 
ipt_MASQUERADE iptable_nat xt_tcpudp bridge iptable_filter bpfilter openvswitch 
nsh nf_conntrack_ipv6 nf_nat_ipv6 nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 
nf_defrag_ipv6 nf_conncount nf_nat nf_conntrack 8021q garp mrp stp llc 
binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) dm_service_time zcommon(PO) 
znvpair(PO) dm_multipath scsi_dh_rdac scsi_dh_emc spl(O) scsi_dh_alua 
ghash_s390 prng aes_s390 des_s390 des_generic sha512_s390 sha256_s390
  [54084.082354]  sha1_s390 sha_common genwqe_card crc_itu_t chsc_sch eadm_sch 
ctcm fsm vfio_ccw vfio_mdev mdev vfio_iommu_type1 vfio sch_fq_codel nfsd 
ib_iser rdma_cm auth_rpcgss nfs_acl iw_cm lockd grace ib_cm iscsi_tcp 
libiscsi_tcp sunrpc libiscsi scsi_transport_iscsi ip_tables x_tables btrfs 
zstd_compress zlib_deflate raid10 raid456 async_raid6_recov async_memcpy 
async_pq async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 linear mlx4_en 
mlx4_ib qeth_l2 ib_core ptp pps_core crc32_vx_s390 mlx4_core devlink zfcp 
scsi_transport_fc qeth qdio ccwgroup dasd_eckd_mod dasd_mod
  [54084.082418] CPU: 3 PID: 217520 Comm: stress-ng-dev Tainted: P  DO  
4.18.0-11-generic #12-Ubuntu
  [54084.082420] Hardware name: IBM 2964 N63 400 (LPAR)
  [54084.082422] Krnl PSW : ed7ed85f 816ddc5c 
(usercopy_abort+0xb2/0xb8)
  [54084.082433]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:0 PM:0 
RI:0 EA:3
  [54084.082436] Krnl GPRS: 0320 0006 0056 
0007
  [54084.082438]0007 7fe0 00a96d3a 

  [54084.082441]03ff 0006 0006 
0002acb4ee00
  [54084.082443]00a96d3a 00914430 00386606 
00038f087a78
  [54084.082449] Krnl Code: 003865fa: c02000390dc4  larl
%r2,aa8182
00386600: c0e5fff2b418  brasl   
%r14,1dce30
   #00386606: a7f40001  brc 
15,386608
   >0038660a: 0707  bcr 0,%r7
0038660c: 0707  bcr 0,%r7
0038660e: 0707  bcr 0,%r7
00386610: c004  brcl0,386610
00386616: ebcff0780024  stmg
%r12,%r15,120(%r15)
  [54084.082484] Call Trace:
  [54084.082487] ([<00386606>] usercopy_abort+0xae/0xb8)
  [54084.082489]  [<00386522>] __check_object_size+0x19a/0x1d0 
  [54084.082492]  [<005c69d8>] sg_scsi_ioctl+0x120/0x3d8 
  [54084.082495]  [<005c75b6>] scsi_cmd_ioctl+0x3ce/0x3d8 
  [54084.082497]  [<005c87d0>] bsg_ioctl+0x118/0x250 
  [54084.082499]  [<003a3c9c>] do_vfs_ioctl+0xac/0x650 
  [54084.082502]  [<003a42cc>] ksys_ioctl+0x8c/0xb8 
  [54084.082504]  [<003a432a>] sys_ioctl+0x32/0x40 
  [54084.082507]  [<008f0dc0>] system_call+0xdc/0x2c8 
  [54084.082508] Last Breaking-Event-Address:
  [54084.082510]  [<00386606>] usercopy_abort+0xae/0xb8
  [54084.082512]  
  [54084.082513] ---[ end trace 5009dd6293a23031 ]---

  ProblemType: Bug
  DistroRelease: Ubuntu 18.10
  Package: linux-image-4.18.0-11-generic 4.18.0-11.12
  ProcVersionSignature: Ubuntu 4.18.0-11.12-generic 4.18.12
  Uname: Linux 4.18.0-11-generic s390x
  NonfreeKernelModules: zfs zunicode zavl icp zcommon znvpair
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay': 'aplay'
  ApportVersion: 2.20.10-0ubuntu13
  Architecture: s390x
  

[Kernel-packages] [Bug 1800609] Re: dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on LPAR

2018-10-30 Thread Colin Ian King
Fix committed to stress-ng:

http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=53e5a98cd58d3b146cf8ffe52540722ff410a5ad

FIBMAP on a block device is plainly wrong, it should only be applied to
a file. The FIBMAP ioctl to a scsi block device is the same as the
SCSI_IOCTL_SEND_COMMAND ioctl that requires a scsi_ioctl_command struct.
As it stands, we are passing an int with stack garbage as this struct
which is passing garbage to the ioctl. The deprecated SCSI ioctl creates
a scsi_request from the garbage and this causes the kernel to correctly
warn with a usercopy_abort that the structure is bogos. Fix this by not
calling FIBMAP on a raw SCSI device.


** Changed in: stress-ng
   Status: New => Fix Committed

** Changed in: stress-ng
   Importance: Undecided => High

** Changed in: stress-ng
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu)
   Status: In Progress => Invalid

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1800609

Title:
  dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on
  LPAR

Status in Stress-ng:
  Fix Committed
Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Invalid

Bug description:
  The dev test on a s390x node (Ubuntu on LPAR) will hang, kernel bug
  could be found in the syslog

  [54084.082277] [ cut here ]
  [54084.082279] kernel BUG at mm/usercopy.c:100!
  [54084.082294] illegal operation: 0001 ilc:1 [#256] SMP 
  [54084.082297] Modules linked in: vhost_net vhost tap cuse dccp_ipv4 dccp 
algif_rng salsa20_generic camellia_generic cast6_generic cast_common 
serpent_generic twofish_generic twofish_common lrw algif_skcipher tgr192 wp512 
rmd320 rmd256 rmd160 rmd128 md4 algif_hash af_alg aufs ip6t_MASQUERADE 
ip6table_nat ip6table_filter ip6_tables xt_comment xt_CHECKSUM iptable_mangle 
ipt_MASQUERADE iptable_nat xt_tcpudp bridge iptable_filter bpfilter openvswitch 
nsh nf_conntrack_ipv6 nf_nat_ipv6 nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 
nf_defrag_ipv6 nf_conncount nf_nat nf_conntrack 8021q garp mrp stp llc 
binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) dm_service_time zcommon(PO) 
znvpair(PO) dm_multipath scsi_dh_rdac scsi_dh_emc spl(O) scsi_dh_alua 
ghash_s390 prng aes_s390 des_s390 des_generic sha512_s390 sha256_s390
  [54084.082354]  sha1_s390 sha_common genwqe_card crc_itu_t chsc_sch eadm_sch 
ctcm fsm vfio_ccw vfio_mdev mdev vfio_iommu_type1 vfio sch_fq_codel nfsd 
ib_iser rdma_cm auth_rpcgss nfs_acl iw_cm lockd grace ib_cm iscsi_tcp 
libiscsi_tcp sunrpc libiscsi scsi_transport_iscsi ip_tables x_tables btrfs 
zstd_compress zlib_deflate raid10 raid456 async_raid6_recov async_memcpy 
async_pq async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 linear mlx4_en 
mlx4_ib qeth_l2 ib_core ptp pps_core crc32_vx_s390 mlx4_core devlink zfcp 
scsi_transport_fc qeth qdio ccwgroup dasd_eckd_mod dasd_mod
  [54084.082418] CPU: 3 PID: 217520 Comm: stress-ng-dev Tainted: P  DO  
4.18.0-11-generic #12-Ubuntu
  [54084.082420] Hardware name: IBM 2964 N63 400 (LPAR)
  [54084.082422] Krnl PSW : ed7ed85f 816ddc5c 
(usercopy_abort+0xb2/0xb8)
  [54084.082433]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:0 PM:0 
RI:0 EA:3
  [54084.082436] Krnl GPRS: 0320 0006 0056 
0007
  [54084.082438]0007 7fe0 00a96d3a 

  [54084.082441]03ff 0006 0006 
0002acb4ee00
  [54084.082443]00a96d3a 00914430 00386606 
00038f087a78
  [54084.082449] Krnl Code: 003865fa: c02000390dc4  larl
%r2,aa8182
00386600: c0e5fff2b418  brasl   
%r14,1dce30
   #00386606: a7f40001  brc 
15,386608
   >0038660a: 0707  bcr 0,%r7
0038660c: 0707  bcr 0,%r7
0038660e: 0707  bcr 0,%r7
00386610: c004  brcl0,386610
00386616: ebcff0780024  stmg
%r12,%r15,120(%r15)
  [54084.082484] Call Trace:
  [54084.082487] ([<00386606>] usercopy_abort+0xae/0xb8)
  [54084.082489]  [<00386522>] __check_object_size+0x19a/0x1d0 
  [54084.082492]  [<005c69d8>] sg_scsi_ioctl+0x120/0x3d8 
  [54084.082495]  [<005c75b6>] scsi_cmd_ioctl+0x3ce/0x3d8 
  [54084.082497]  [<005c87d0>] bsg_ioctl+0x118/0x250 
  [54084.082499]  [<003a3c9c>] do_vfs_ioctl+0xac/0x650 
  [54084.082502]  [<003a42cc>] ksys_ioctl+0x8c/0xb8 
  [54084.082504]  [<003a432a

[Kernel-packages] [Bug 1800609] Re: dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on LPAR

2018-10-30 Thread Colin Ian King
** Changed in: linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu)
   Importance: Undecided => Medium

** Changed in: linux (Ubuntu)
   Status: Incomplete => In Progress

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1800609

Title:
  dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on
  LPAR

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  The dev test on a s390x node (Ubuntu on LPAR) will hang, kernel bug
  could be found in the syslog

  [54084.082277] [ cut here ]
  [54084.082279] kernel BUG at mm/usercopy.c:100!
  [54084.082294] illegal operation: 0001 ilc:1 [#256] SMP 
  [54084.082297] Modules linked in: vhost_net vhost tap cuse dccp_ipv4 dccp 
algif_rng salsa20_generic camellia_generic cast6_generic cast_common 
serpent_generic twofish_generic twofish_common lrw algif_skcipher tgr192 wp512 
rmd320 rmd256 rmd160 rmd128 md4 algif_hash af_alg aufs ip6t_MASQUERADE 
ip6table_nat ip6table_filter ip6_tables xt_comment xt_CHECKSUM iptable_mangle 
ipt_MASQUERADE iptable_nat xt_tcpudp bridge iptable_filter bpfilter openvswitch 
nsh nf_conntrack_ipv6 nf_nat_ipv6 nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 
nf_defrag_ipv6 nf_conncount nf_nat nf_conntrack 8021q garp mrp stp llc 
binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) dm_service_time zcommon(PO) 
znvpair(PO) dm_multipath scsi_dh_rdac scsi_dh_emc spl(O) scsi_dh_alua 
ghash_s390 prng aes_s390 des_s390 des_generic sha512_s390 sha256_s390
  [54084.082354]  sha1_s390 sha_common genwqe_card crc_itu_t chsc_sch eadm_sch 
ctcm fsm vfio_ccw vfio_mdev mdev vfio_iommu_type1 vfio sch_fq_codel nfsd 
ib_iser rdma_cm auth_rpcgss nfs_acl iw_cm lockd grace ib_cm iscsi_tcp 
libiscsi_tcp sunrpc libiscsi scsi_transport_iscsi ip_tables x_tables btrfs 
zstd_compress zlib_deflate raid10 raid456 async_raid6_recov async_memcpy 
async_pq async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 linear mlx4_en 
mlx4_ib qeth_l2 ib_core ptp pps_core crc32_vx_s390 mlx4_core devlink zfcp 
scsi_transport_fc qeth qdio ccwgroup dasd_eckd_mod dasd_mod
  [54084.082418] CPU: 3 PID: 217520 Comm: stress-ng-dev Tainted: P  DO  
4.18.0-11-generic #12-Ubuntu
  [54084.082420] Hardware name: IBM 2964 N63 400 (LPAR)
  [54084.082422] Krnl PSW : ed7ed85f 816ddc5c 
(usercopy_abort+0xb2/0xb8)
  [54084.082433]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:0 PM:0 
RI:0 EA:3
  [54084.082436] Krnl GPRS: 0320 0006 0056 
0007
  [54084.082438]0007 7fe0 00a96d3a 

  [54084.082441]03ff 0006 0006 
0002acb4ee00
  [54084.082443]00a96d3a 00914430 00386606 
00038f087a78
  [54084.082449] Krnl Code: 003865fa: c02000390dc4  larl
%r2,aa8182
00386600: c0e5fff2b418  brasl   
%r14,1dce30
   #00386606: a7f40001  brc 
15,386608
   >0038660a: 0707  bcr 0,%r7
0038660c: 0707  bcr 0,%r7
0038660e: 0707  bcr 0,%r7
00386610: c004  brcl0,386610
00386616: ebcff0780024  stmg
%r12,%r15,120(%r15)
  [54084.082484] Call Trace:
  [54084.082487] ([<00386606>] usercopy_abort+0xae/0xb8)
  [54084.082489]  [<00386522>] __check_object_size+0x19a/0x1d0 
  [54084.082492]  [<005c69d8>] sg_scsi_ioctl+0x120/0x3d8 
  [54084.082495]  [<005c75b6>] scsi_cmd_ioctl+0x3ce/0x3d8 
  [54084.082497]  [<005c87d0>] bsg_ioctl+0x118/0x250 
  [54084.082499]  [<003a3c9c>] do_vfs_ioctl+0xac/0x650 
  [54084.082502]  [<003a42cc>] ksys_ioctl+0x8c/0xb8 
  [54084.082504]  [<003a432a>] sys_ioctl+0x32/0x40 
  [54084.082507]  [<008f0dc0>] system_call+0xdc/0x2c8 
  [54084.082508] Last Breaking-Event-Address:
  [54084.082510]  [<00386606>] usercopy_abort+0xae/0xb8
  [54084.082512]  
  [54084.082513] ---[ end trace 5009dd6293a23031 ]---

  ProblemType: Bug
  DistroRelease: Ubuntu 18.10
  Package: linux-image-4.18.0-11-generic 4.18.0-11.12
  ProcVersionSignature: Ubuntu 4.18.0-11.12-generic 4.18.12
  Uname: Linux 4.18.0-11-generic s390x
  NonfreeKernelModules: zfs zunicode zavl icp zcommon znvpair
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No suc

[Kernel-packages] [Bug 1800609] Re: dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on LPAR

2018-10-30 Thread Colin Ian King
Occurs when performing the FIBMAP ioctl on /dev/bsg/rport-1:0-0

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1800609

Title:
  dev test in ubuntu_stress_smoke_tests will hang on s390x Ubuntu on
  LPAR

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  The dev test on a s390x node (Ubuntu on LPAR) will hang, kernel bug
  could be found in the syslog

  [54084.082277] [ cut here ]
  [54084.082279] kernel BUG at mm/usercopy.c:100!
  [54084.082294] illegal operation: 0001 ilc:1 [#256] SMP 
  [54084.082297] Modules linked in: vhost_net vhost tap cuse dccp_ipv4 dccp 
algif_rng salsa20_generic camellia_generic cast6_generic cast_common 
serpent_generic twofish_generic twofish_common lrw algif_skcipher tgr192 wp512 
rmd320 rmd256 rmd160 rmd128 md4 algif_hash af_alg aufs ip6t_MASQUERADE 
ip6table_nat ip6table_filter ip6_tables xt_comment xt_CHECKSUM iptable_mangle 
ipt_MASQUERADE iptable_nat xt_tcpudp bridge iptable_filter bpfilter openvswitch 
nsh nf_conntrack_ipv6 nf_nat_ipv6 nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 
nf_defrag_ipv6 nf_conncount nf_nat nf_conntrack 8021q garp mrp stp llc 
binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) dm_service_time zcommon(PO) 
znvpair(PO) dm_multipath scsi_dh_rdac scsi_dh_emc spl(O) scsi_dh_alua 
ghash_s390 prng aes_s390 des_s390 des_generic sha512_s390 sha256_s390
  [54084.082354]  sha1_s390 sha_common genwqe_card crc_itu_t chsc_sch eadm_sch 
ctcm fsm vfio_ccw vfio_mdev mdev vfio_iommu_type1 vfio sch_fq_codel nfsd 
ib_iser rdma_cm auth_rpcgss nfs_acl iw_cm lockd grace ib_cm iscsi_tcp 
libiscsi_tcp sunrpc libiscsi scsi_transport_iscsi ip_tables x_tables btrfs 
zstd_compress zlib_deflate raid10 raid456 async_raid6_recov async_memcpy 
async_pq async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 linear mlx4_en 
mlx4_ib qeth_l2 ib_core ptp pps_core crc32_vx_s390 mlx4_core devlink zfcp 
scsi_transport_fc qeth qdio ccwgroup dasd_eckd_mod dasd_mod
  [54084.082418] CPU: 3 PID: 217520 Comm: stress-ng-dev Tainted: P  DO  
4.18.0-11-generic #12-Ubuntu
  [54084.082420] Hardware name: IBM 2964 N63 400 (LPAR)
  [54084.082422] Krnl PSW : ed7ed85f 816ddc5c 
(usercopy_abort+0xb2/0xb8)
  [54084.082433]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:0 PM:0 
RI:0 EA:3
  [54084.082436] Krnl GPRS: 0320 0006 0056 
0007
  [54084.082438]0007 7fe0 00a96d3a 

  [54084.082441]03ff 0006 0006 
0002acb4ee00
  [54084.082443]00a96d3a 00914430 00386606 
00038f087a78
  [54084.082449] Krnl Code: 003865fa: c02000390dc4  larl
%r2,aa8182
00386600: c0e5fff2b418  brasl   
%r14,1dce30
   #00386606: a7f40001  brc 
15,386608
   >0038660a: 0707  bcr 0,%r7
0038660c: 0707  bcr 0,%r7
0038660e: 0707  bcr 0,%r7
00386610: c004  brcl0,386610
00386616: ebcff0780024  stmg
%r12,%r15,120(%r15)
  [54084.082484] Call Trace:
  [54084.082487] ([<00386606>] usercopy_abort+0xae/0xb8)
  [54084.082489]  [<00386522>] __check_object_size+0x19a/0x1d0 
  [54084.082492]  [<005c69d8>] sg_scsi_ioctl+0x120/0x3d8 
  [54084.082495]  [<005c75b6>] scsi_cmd_ioctl+0x3ce/0x3d8 
  [54084.082497]  [<005c87d0>] bsg_ioctl+0x118/0x250 
  [54084.082499]  [<003a3c9c>] do_vfs_ioctl+0xac/0x650 
  [54084.082502]  [<003a42cc>] ksys_ioctl+0x8c/0xb8 
  [54084.082504]  [<003a432a>] sys_ioctl+0x32/0x40 
  [54084.082507]  [<008f0dc0>] system_call+0xdc/0x2c8 
  [54084.082508] Last Breaking-Event-Address:
  [54084.082510]  [<00386606>] usercopy_abort+0xae/0xb8
  [54084.082512]  
  [54084.082513] ---[ end trace 5009dd6293a23031 ]---

  ProblemType: Bug
  DistroRelease: Ubuntu 18.10
  Package: linux-image-4.18.0-11-generic 4.18.0-11.12
  ProcVersionSignature: Ubuntu 4.18.0-11.12-generic 4.18.12
  Uname: Linux 4.18.0-11-generic s390x
  NonfreeKernelModules: zfs zunicode zavl icp zcommon znvpair
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay': 'aplay'
  ApportVersion: 2.20.10-0ubuntu13
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord': 
'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with 

[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-10-30 Thread Colin Ian King
Tested with linux 4.4.0-139-generic #165 and zfsutils 0.6.5.6-0ubuntu26
- marking as verified for xenial

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  Fix Committed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, Xenial, Bionic ==

  Bug:

  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

  == Fix ==

  Upstream zfs commit 4ceb8dd6fdfd ("Fix 'zpool create -t '")

  == Test ==

  Without the fix, creating a pool with -t will show an error message
  such as:

  "cannot create 'rpool': no such pool or dataset" and will return an
  error number even though it successfully created the pool. With the
  fix, the error is not shown and it exits with 0.

  E.g.

  pool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1 
  cannot create 'rpool': no such pool or dataset

  == Regression Potential ==

  Small. This is an upstream fix that is know to fix the error. There
  are no known subsequent changes to the fix hence suggesting it is a
  good fix. The change is very limited and testing shows it does not
  break any of the kernel team's ZFS regression testing.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1793464] Re: Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

2018-10-25 Thread Colin Ian King
Tested against -proposed kernel 4.4.0-139-generic, nbd tests no longer
fail. Marking as verified.

** Tags removed: verification-needed-xenial
** Tags added: verification-done-xenial

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1793464

Title:
  Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress
Status in linux source package in Xenial:
  Fix Committed

Bug description:
  == SRU Justification ==

  When running the Ubuntu nbd autotest regression test we trip a hang
  and then a little later a panic message.  There are two upstream
  fixes required as this is actually two issues in one. One fix is to 
  not to shutdown the sock when IRQs are disable and a second to fix is
  to race in the nbd ioctl.

  == Fix ==

  Upstream commits:

  23272a6754b81ff6503e09c743bb4ceeeab39997
nbd: Remove signal usage

  1f7b5cf1be4351e60cf8ae7aab976503dd73c5f8
nbd: Timeouts are not user requested disconnects

  0e4f0f6f63d3416a9e529d99febfe98545427b81
nbd: Cleanup reset of nbd and bdev after a disconnect

  c261189862c6f65117eb3b1748622a08ef49c262
nbd: don't shutdown sock with irq's disabled

  97240963eb308d8d21a89c0459822f7ea98463b4
nbd: fix race in ioctl

  The first 3 patches are prerequisites required for the latter two fixes to 
apply and work correctly.  Most of these backports are minor patch wiggles
  required because later patches have been applied to the driver in earlier 
fixes to this driver.
 

  == Regression Potential ==

  These fixes just touch nbd, so the regression potential is just
  limited to this. Secondly, we are pulling in upstream fixes that exist
  in Bionic and Cosmic kernels, so these are tried and tested fixes.

  == Test Case ==

1. Deploy a node with 4.4 Xenial
2. Run the ubuntu_nbd_smoke_test

  Without the fix, we get hang/crashes.  With the fix one can run this test
  multiple times without any issues at all.

  
  

  This issue can be reproduced on AMD64 KVM //bare-metal node, s390x
  zKVM node

  The test itself will pass, but the system will hang after a few
  second.

  Steps:
    1. Deploy a node with 4.4 Xenial
    2. Run the ubuntu_nbd_smoke_test

  If you have access to the console, you will see that this system
  actually bumped into a kernel panic:

   Unable to handle kernel pointer dereference in virtual kernel address space
   failing address: 03ff802c1000 TEID: 03ff802c1803
   Fault in home space mode while using kernel ASCE.
   Log here (s390x KVM): https://pastebin.ubuntu.com/p/dNmtvbGjmz/

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-136-generic 4.4.0-136.162
  ProcVersionSignature: Ubuntu 4.4.0-136.162-generic 4.4.144
  Uname: Linux 4.4.0-136-generic s390x
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  Date: Thu Sep 20 03:46:00 2018
  HibernationDevice: RESUME=UUID=ca468a9c-9563-442c-85c6-6055e800a66e
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig'
  Lspci:

  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:

  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=UUID=b65b756a-ba4e-4c53-aa32-0db2bdb50bb3 
crashkernel=196M
  RelatedPackageVersions:
   linux-restricted-modules-4.4.0-136-generic N/A
   linux-backports-modules-4.4.0-136-generic  N/A
   linux-firmware 1.157.20
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1793464/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-10-25 Thread Colin Ian King
Tested with linux 4.4.0-193 and zfsutils 0.6.5.6-0ubuntu25 - marking as
verified for xenial

** Tags removed: verification-needed-xenial
** Tags added: verification-done-xenial

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  New
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, Xenial, Bionic ==

  Bug:

  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

  == Fix ==

  Upstream zfs commit 4ceb8dd6fdfd ("Fix 'zpool create -t '")

  == Test ==

  Without the fix, creating a pool with -t will show an error message
  such as:

  "cannot create 'rpool': no such pool or dataset" and will return an
  error number even though it successfully created the pool. With the
  fix, the error is not shown and it exits with 0.

  E.g.

  pool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1 
  cannot create 'rpool': no such pool or dataset

  == Regression Potential ==

  Small. This is an upstream fix that is know to fix the error. There
  are no known subsequent changes to the fix hence suggesting it is a
  good fix. The change is very limited and testing shows it does not
  break any of the kernel team's ZFS regression testing.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1797327] Re: getrandom02 in ubuntu_ltp_syscalls failed with X-kvm

2018-10-11 Thread Colin Ian King
I suspect that the entropy pool ran out of bits

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1797327

Title:
  getrandom02 in ubuntu_ltp_syscalls failed with X-kvm

Status in ubuntu-kernel-tests:
  New
Status in linux-kvm package in Ubuntu:
  New

Bug description:
  This test will time out with X-KVM kernel on a KVM node.

  <<>>
  tag=getrandom02 stime=1539245024
  cmdline="getrandom02"
  contacts=""
  analysis=exit
  <<>>
  tst_test.c:1072: INFO: Timeout per run is 0h 05m 00s
  getrandom02.c:72: PASS: getrandom returned 256
  getrandom02.c:72: PASS: getrandom returned 6
  getrandom02.c:72: PASS: getrandom returned 256
  Test timeouted, sending SIGKILL!
  tst_test.c:1112: INFO: If you are running on slow machine, try exporting 
LTP_TIMEOUT_MUL > 1
  tst_test.c:1113: BROK: Test killed! (timeout?)

  Summary:
  passed   3
  failed   0
  skipped  0
  warnings 0

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-1036-kvm 4.4.0-1036.42
  ProcVersionSignature: User Name 4.4.0-1036.42-kvm 4.4.155
  Uname: Linux 4.4.0-1036-kvm x86_64
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: amd64
  Date: Thu Oct 11 08:05:56 2018
  SourcePackage: linux-kvm
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1797327/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1793464] Re: Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

2018-10-11 Thread Colin Ian King
** Description changed:

+ == SRU Justification ==
+ 
+ When running the Ubuntu nbd autotest regression test we trip a hang
+ and then a little later a panic message.  There are two upstream
+ fixes required as this is actually two issues in one. One fix is to 
+ not to shutdown the sock when IRQs are disable and a second to fix is
+ to race in the nbd ioctl.
+ 
+ == Fix ==
+ 
+ Upstream commits:
+ 
+ 23272a6754b81ff6503e09c743bb4ceeeab39997
+   nbd: Remove signal usage
+ 
+ 1f7b5cf1be4351e60cf8ae7aab976503dd73c5f8
+   nbd: Timeouts are not user requested disconnects
+ 
+ 0e4f0f6f63d3416a9e529d99febfe98545427b81
+   nbd: Cleanup reset of nbd and bdev after a disconnect
+ 
+ c261189862c6f65117eb3b1748622a08ef49c262
+   nbd: don't shutdown sock with irq's disabled
+ 
+ 97240963eb308d8d21a89c0459822f7ea98463b4
+   nbd: fix race in ioctl
+ 
+ The first 3 patches are prerequisites required for the latter two fixes to 
apply and work correctly.  Most of these backports are minor patch wiggles
+ required because later patches have been applied to the driver in earlier 
fixes to this driver.
+
+ 
+ == Regression Potential ==
+ 
+ These fixes just touch nbd, so the regression potential is just limited
+ to this. Secondly, we are pulling in upstream fixes that exist in Bionic
+ and Cosmic kernels, so these are tried and tested fixes.
+ 
+ == Test Case ==
+ 
+   1. Deploy a node with 4.4 Xenial
+   2. Run the ubuntu_nbd_smoke_test
+ 
+ Without the fix, we get hang/crashes.  With the fix one can run this test
+ multiple times without any issues at all.
+ 
+ 
+ 
+ 
  This issue can be reproduced on AMD64 KVM //bare-metal node, s390x zKVM
  node
  
  The test itself will pass, but the system will hang after a few second.
  
  Steps:
    1. Deploy a node with 4.4 Xenial
    2. Run the ubuntu_nbd_smoke_test
  
  If you have access to the console, you will see that this system
  actually bumped into a kernel panic:
  
   Unable to handle kernel pointer dereference in virtual kernel address space
   failing address: 03ff802c1000 TEID: 03ff802c1803
   Fault in home space mode while using kernel ASCE.
   Log here (s390x KVM): https://pastebin.ubuntu.com/p/dNmtvbGjmz/
  
  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-136-generic 4.4.0-136.162
  ProcVersionSignature: Ubuntu 4.4.0-136.162-generic 4.4.144
  Uname: Linux 4.4.0-136-generic s390x
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  Date: Thu Sep 20 03:46:00 2018
  HibernationDevice: RESUME=UUID=ca468a9c-9563-442c-85c6-6055e800a66e
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig'
  Lspci:
  
  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:
  
  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=UUID=b65b756a-ba4e-4c53-aa32-0db2bdb50bb3 
crashkernel=196M
  RelatedPackageVersions:
   linux-restricted-modules-4.4.0-136-generic N/A
   linux-backports-modules-4.4.0-136-generic  N/A
   linux-firmware 1.157.20
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  UpgradeStatus: No upgrade log present (probably fresh install)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1793464

Title:
  Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  == SRU Justification ==

  When running the Ubuntu nbd autotest regression test we trip a hang
  and then a little later a panic message.  There are two upstream
  fixes required as this is actually two issues in one. One fix is to 
  not to shutdown the sock when IRQs are disable and a second to fix is
  to race in the nbd ioctl.

  == Fix ==

  Upstream commits:

  23272a6754b81ff6503e09c743bb4ceeeab39997
nbd: Remove signal usage

  1f7b5cf1be4351e60cf8ae7aab976503dd73c5f8
nbd: Timeouts are not user requested disconnects

  0e4f0f6f63d3416a9e529d99febfe98545427b81
nbd: Cleanup reset of nbd and bdev after a disconnect

  c261189862c6f65117eb3b1748622a08ef49c262
nbd: don't shutdown sock with irq's disabled

  97240963eb308d8d21a89c0459822f7ea98463b4
nbd: fix race in ioctl

  The first 3 patches are prerequisites required for the latter two fixes to 
apply and work correctly.  Most of these backports are minor patch wiggles
  required because later patches have been applied to the 

[Kernel-packages] [Bug 1796250] Re: sysfs test ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-11 Thread Colin Ian King
** Changed in: linux (Ubuntu)
   Status: Fix Committed => Won't Fix

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  sysfs test ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  Fix Released
Status in linux package in Ubuntu:
  Won't Fix

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 DEBUG| [stdout] [ 1422.281009]  [] ? 

[Kernel-packages] [Bug 1796250] Re: sysfs test ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-09 Thread Colin Ian King
The number of prerequisite fixes before we can apply the fix
748935eeb72c343 makes this a rather overly involved fix.  I doubt it
will be SRU'able.  Since this *only* occurs when accessing two TPM
related interfaces in a fast multiple multi-threaded race on a few
specific x86 devices as root makes this a very risky set of backport
changes for such a corner case.

For now, I'm going to force stress-ng sysfs test to skip the tpm files
for older kernels to workaround this issue. (Ugh).

Workaround committed in stress-ng:

http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=c7fcb4112b97188c8fcba6138b29b5c5a82938ea

** Changed in: linux (Ubuntu)
   Status: In Progress => Fix Committed

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  sysfs test ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Fix Committed

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 

[Kernel-packages] [Bug 1796542] Re: Silent data corruption in Linux kernel 4.15

2018-10-09 Thread Colin Ian King
** Changed in: linux (Ubuntu Bionic)
   Status: Triaged => In Progress

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796542

Title:
  Silent data corruption in Linux kernel 4.15

Status in linux package in Ubuntu:
  Fix Released
Status in linux source package in Bionic:
  In Progress

Bug description:
  == SRU Justification [BIONIC] ==

  A silent data corruption was introduced in v4.10-rc1 with commit
  72ecad22d9f198aafee64218512e02ffa7818671 and was fixed in v4.18-rc7
  with commit 17d51b10d7773e4618bcac64648f30f12d4078fb. It affects
  users of O_DIRECT, in our case a KVM virtual machine with drives
  which use qemu's "cache=none" option.

  == Fix ==

  Upstream commits:

  0aa69fd32a5f766e997ca8ab4723c5a1146efa8b
    block: add a lower-level bio_add_page interface

  b403ea2404889e1227812fa9657667a1deb9c694
    block: bio_iov_iter_get_pages: fix size of last iovec

  9362dd1109f87a9d0a798fbc890cb339c171ed35
    blkdev: __blkdev_direct_IO_simple: fix leak in error case

  17d51b10d7773e4618bcac64648f30f12d4078fb
    block: bio_iov_iter_get_pages: pin more pages for multi-segment IOs

  The first 3 patches are required for a clean application of the final
  patch that actually addresses the problem with a fix to this known
  issue.

  == Regression Potential ==

  This touches the block layer, so there is risk potential in data
  corruption. The fixes have several weeks in the upstream kernel and
  so far, I see no subsequent fixes required.

  == Test Case ==

  Build the program listed below [1]
  kudos to Jan Kara, and run with:

  dd if=/dev/zero if=loop.img bs=1M count=2048
  sudo losetup /dev/loop0 loop.img

  ./blkdev-dio-test /dev/loop0 0 &
  ./blkdev-dio-test /dev/loop0 2048 &

  Without the fix, ones lost writes fairly soon.  Without the fix, this
  runs without any losy write messages.

  blkdev-dio-test.c:

  #define _GNU_SOURCE
  #include 
  #include 
  #include 
  #include 
  #include 
  #include 

  #define PAGE_SIZE 4096
  #define SECT_SIZE 512
  #define BUF_OFF (2*SECT_SIZE)

  int main(int argc, char **argv)
  {
   int fd = open(argv[1], O_RDWR | O_DIRECT);
   int ret;
   char *buf;
   loff_t off;
   struct iovec iov[2];
   unsigned int seq;

   if (fd < 0) {
    perror("open");
    return 1;
   }

   off = strtol(argv[2], NULL, 10);

   buf = aligned_alloc(PAGE_SIZE, PAGE_SIZE);

   iov[0].iov_base = buf;
   iov[0].iov_len = SECT_SIZE;
   iov[1].iov_base = buf + BUF_OFF;
   iov[1].iov_len = SECT_SIZE;

   seq = 0;
   memset(buf, 0, PAGE_SIZE);
   while (1) {
    *(unsigned int *)buf = seq;
    *(unsigned int *)(buf + BUF_OFF) = seq;
    ret = pwritev(fd, iov, 2, off);
    if (ret < 0) {
     perror("pwritev");
     return 1;
    }
    if (ret != 2*SECT_SIZE) {
     fprintf(stderr, "Short pwritev: %d\n", ret);
     return 1;
    }
    ret = pread(fd, buf, PAGE_SIZE, off);
    if (ret < 0) {
     perror("pread");
     return 1;
    }
    if (ret != PAGE_SIZE) {
     fprintf(stderr, "Short read: %d\n", ret);
     return 1;
    }
    if (*(unsigned int *)buf != seq ||
    *(unsigned int *)(buf + SECT_SIZE) != seq) {
     printf("Lost write %u: %u %u\n", seq, *(unsigned int *)buf, *(unsigned int 
*)(buf + SECT_SIZE));
     return 1;
    }
    seq++;
   }

   return 0;
  }

  References:
  [1] https://www.spinics.net/lists/linux-block/msg28507.html

  

  TLDR: commit 72ecad22d9f198aafee64218512e02ffa7818671 (in v4.10)
  introduced silent data corruption for O_DIRECT uses, it's fixed in
  17d51b10d7773e4618bcac64648f30f12d4078fb (in v4.18)

  A silent data corruption was introduced in v4.10-rc1 with commit
  72ecad22d9f198aafee64218512e02ffa7818671 and was fixed in v4.18-rc7
  with commit 17d51b10d7773e4618bcac64648f30f12d4078fb. It affects users
  of O_DIRECT, in our case a KVM virtual machine with drives which use
  qemu's "cache=none" option.

  This is the commit which fixes the issue:
  -
  commit 17d51b10d7773e4618bcac64648f30f12d4078fb
  Author: Martin Wilck 
  Date:   Wed Jul 25 23:15:09 2018 +0200

  block: bio_iov_iter_get_pages: pin more pages for multi-segment
  IOs

  bio_iov_iter_get_pages() currently only adds pages for the next non-zero
  segment from the iov_iter to the bio. That's suboptimal for callers,
  which typically try to pin as many pages as fit into the bio. This patch
  converts the current bio_iov_iter_get_pages() into a static helper, and
  introduces a new helper that allocates as many pages as

   1) fit into the bio,
   2) are present in the iov_iter,
   3) and can be pinned by MM.

  Error is returned only if zero pages could be pinned. Because of 3), a
  zero return value doesn't necessarily mean all pages have been pinned.
  Callers that have to pin every page in the iov_iter must still call this
  function in a loop (this is currently the case).

[Kernel-packages] [Bug 1796542] Re: Silent corruption in Linux kernel 4.15

2018-10-09 Thread Colin Ian King
OK, I've explored the fixes, it addresses the issue and I can't
reproduce the problem with the fix testing with the provided reproducer
as referenced in comment #3

** Description changed:

+ == SRU Justification ==
+ 
+ A silent data corruption was introduced in v4.10-rc1 with commit
+ 72ecad22d9f198aafee64218512e02ffa7818671 and was fixed in v4.18-rc7
+ with commit 17d51b10d7773e4618bcac64648f30f12d4078fb. It affects
+ users of O_DIRECT, in our case a KVM virtual machine with drives
+ which use qemu's "cache=none" option.
+ 
+ == Fix ==
+ 
+ Upstream commits:
+ 
+ 0aa69fd32a5f766e997ca8ab4723c5a1146efa8b
+   block: add a lower-level bio_add_page interface
+ 
+ b403ea2404889e1227812fa9657667a1deb9c694
+   block: bio_iov_iter_get_pages: fix size of last iovec
+ 
+ 9362dd1109f87a9d0a798fbc890cb339c171ed35 
+   blkdev: __blkdev_direct_IO_simple: fix leak in error case
+ 
+ 17d51b10d7773e4618bcac64648f30f12d4078fb
+   block: bio_iov_iter_get_pages: pin more pages for multi-segment IOs
+ 
+ The first 3 patches are required for a clean application of the final
+ patch that actually addresses the problem with a fix to this known
+ issue.
+ 
+ == Regression Potential ==
+ 
+ This touches the block layer, so there is risk potential in data
+ corruption. The fixes have several weeks in the upstream kernel and
+ so far, I see no subsequent fixes required.
+ 
+ == Test Case ==
+ 
+ Build the program listed below [1]
+ kudos to Jan Kara, and run with:
+ 
+ dd if=/dev/zero if=loop.img bs=1M count=2048
+ sudo losetup /dev/loop0 loop.img
+ 
+ ./blkdev-dio-test /dev/loop0 0 &
+ ./blkdev-dio-test /dev/loop0 2048 &
+ 
+ Without the fix, ones lost writes fairly soon.  Without the fix, this
+ runs without any losy write messages.
+ 
+ blkdev-dio-test.c:
+ 
+ #define _GNU_SOURCE
+ #include 
+ #include 
+ #include 
+ #include 
+ #include 
+ #include 
+ 
+ #define PAGE_SIZE 4096
+ #define SECT_SIZE 512
+ #define BUF_OFF (2*SECT_SIZE)
+ 
+ int main(int argc, char **argv)
+ {
+   int fd = open(argv[1], O_RDWR | O_DIRECT);
+   int ret;
+   char *buf;
+   loff_t off;
+   struct iovec iov[2];
+   unsigned int seq;
+ 
+   if (fd < 0) {
+   perror("open");
+   return 1;
+   }
+ 
+ off = strtol(argv[2], NULL, 10);
+ 
+ buf = aligned_alloc(PAGE_SIZE, PAGE_SIZE);
+ 
+   iov[0].iov_base = buf;
+   iov[0].iov_len = SECT_SIZE;
+   iov[1].iov_base = buf + BUF_OFF;
+   iov[1].iov_len = SECT_SIZE;
+ 
+   seq = 0;
+   memset(buf, 0, PAGE_SIZE);
+   while (1) {
+   *(unsigned int *)buf = seq;
+   *(unsigned int *)(buf + BUF_OFF) = seq;
+   ret = pwritev(fd, iov, 2, off);
+   if (ret < 0) {
+   perror("pwritev");
+   return 1;
+   }
+   if (ret != 2*SECT_SIZE) {
+   fprintf(stderr, "Short pwritev: %d\n", ret);
+   return 1;
+   }
+   ret = pread(fd, buf, PAGE_SIZE, off);
+   if (ret < 0) {
+   perror("pread");
+   return 1;
+   }
+   if (ret != PAGE_SIZE) {
+   fprintf(stderr, "Short read: %d\n", ret);
+   return 1;
+   }
+   if (*(unsigned int *)buf != seq ||
+   *(unsigned int *)(buf + SECT_SIZE) != seq) {
+   printf("Lost write %u: %u %u\n", seq, *(unsigned int 
*)buf, *(unsigned int *)(buf + SECT_SIZE));
+   return 1;
+   }
+   seq++;
+   }
+ 
+   return 0;
+ }
+ 
+ References:
+ [1] https://www.spinics.net/lists/linux-block/msg28507.html
+ 
+ 
+ 
  TLDR: commit 72ecad22d9f198aafee64218512e02ffa7818671 (in v4.10)
  introduced silent data corruption for O_DIRECT uses, it's fixed in
  17d51b10d7773e4618bcac64648f30f12d4078fb (in v4.18)
  
  A silent data corruption was introduced in v4.10-rc1 with commit
  72ecad22d9f198aafee64218512e02ffa7818671 and was fixed in v4.18-rc7 with
  commit 17d51b10d7773e4618bcac64648f30f12d4078fb. It affects users of
  O_DIRECT, in our case a KVM virtual machine with drives which use qemu's
  "cache=none" option.
  
  This is the commit which fixes the issue:
  -
  commit 17d51b10d7773e4618bcac64648f30f12d4078fb
  Author: Martin Wilck 
  Date:   Wed Jul 25 23:15:09 2018 +0200
  
- block: bio_iov_iter_get_pages: pin more pages for multi-segment IOs
- 
- bio_iov_iter_get_pages() currently only adds pages for the next non-zero
- segment from the iov_iter to the bio. That's suboptimal for callers,
- which typically try to pin as many pages as fit into the bio. This patch
- converts the current bio_iov_iter_get_pages() into a static helper, and
- introduces a new helper that allocates as many pages as
- 
-  1) fit into the bio,
-  2) are present in the 

[Kernel-packages] [Bug 1757057] Re: sysfs test in ubuntu_stress_smoke will induce kernel oops on X-HWE ppc64le

2018-10-09 Thread Colin Ian King
I'd close it, artful is EOL

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1757057

Title:
  sysfs test in ubuntu_stress_smoke will induce kernel oops on X-HWE
  ppc64le

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  Confirmed
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Proposed kernel 4.13.0-38.43~16.04.1 on ppc64le

  sysfs FAILED (kernel oopsed)
  [ 891.312202] WARNING! power/level is deprecated; use power/control instead
  [ 892.138260] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.138627] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.138976] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.139303] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.139610] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 255
  [ 892.141571] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.141697] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142003] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142319] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142616] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142903] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 87
  [ 892.148239] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.148583] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.149406] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.149530] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 255
  [ 892.149855] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.150161] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.150467] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.151980] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.152255] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.152558] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.152980] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.153545] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.153908] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.154197] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.154502] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.154813] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.155110] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.155407] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.162052] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.162406] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.162708] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 85
  [ 892.163007] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.163304] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.163612] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.163912] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.164505] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.164816] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.165118] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.165413] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.165717] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166004] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166309] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166613] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166908] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.167212] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.167524] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.167803] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 22
  [ 892.168128] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.168439] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.168732] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169038] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169335] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169630] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169934] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.170235] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176000] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176350] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176643] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176939] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.177258] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.177546] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.177962] 

[Kernel-packages] [Bug 1796250] Re: sysfs test ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-09 Thread Colin Ian King
Bisected, fixed with upstream commit:

>From 748935eeb72c34368ab514a2bfdf75161768cec0 Mon Sep 17 00:00:00 2001
From: Nayna Jain 
Date: Mon, 14 Nov 2016 05:00:52 -0500
Subject: [PATCH] tpm: have event log use the tpm_chip

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  sysfs test ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  

[Kernel-packages] [Bug 1796250] Re: sysfs test ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-08 Thread Colin Ian King
repeated course bisect, now got something more reasonable:

4.8 fail
4.9 fail
4.10 - ok
4.11 - ok
4.12 - ok
4.14 - ok 
4.18 - ok

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  sysfs test ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 

[Kernel-packages] [Bug 1796250] Re: sysfs test ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-08 Thread Colin Ian King
Course bisect, fixed between 4.8 and 4.9

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  sysfs test ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 DEBUG| [stdout] [ 1422.281009]  [] ? 
find_vma+0x68/0x70
  16:50:05 DEBUG| [stdout] [ 

[Kernel-packages] [Bug 1793464] Re: Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

2018-10-08 Thread Colin Ian King
Course sanity check with mainline kernels:

4.2 OK
4.3 hangs
4.4 hangs
4.5 hang
4.6 - 4.8 no hang, but dumps message (see below)
4.9 OK

Oct  8 11:04:03 ubuntu kernel: [   31.788232] block nbd0: NBD_DISCONNECT
Oct  8 11:04:03 ubuntu kernel: [   31.788286] block nbd0: shutting down socket
Oct  8 11:04:03 ubuntu kernel: [   31.788290] [ cut here 
]
Oct  8 11:04:03 ubuntu kernel: [   31.788299] WARNING: CPU: 0 PID: 1807 at 
/home/kernel/COD/linux/kernel/softirq.c:150 __local_bh_enable_ip+0x6b/0x80
Oct  8 11:04:03 ubuntu kernel: [   31.788301] Modules linked in: squashfs loop 
fuse nbd nls_iso8859_1 vfat fat snd_hda_codec_generic snd_hda_intel 
snd_hda_codec ppdev snd_hda_core virtio_console snd_hwdep virtio_balloon 
snd_pcm joydev input_leds efi_pstore led_class snd_timer efivars serio_raw snd 
i2c_piix4 soundcore acpi_cpufreq parport_pc 8250_fintek processor parport 
qemu_fw_cfg mac_hid ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core ib_addr 
configfs iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi efivarfs autofs4 
ext4 crc16 jbd2 mbcache btrfs raid10 raid456 async_raid6_recov async_memcpy 
async_pq async_xor async_tx xor hid_generic usbhid hid raid6_pq libcrc32c 
crc32c_generic raid1 raid0 multipath linear md_mod ata_generic virtio_net 
virtio_blk pata_acpi qxl ttm drm_kms_helper syscopyarea crct10dif_pclmul 
crc32_pclmul sysfillrect sysimgblt crc32c_intel fb_sys_fops ghash_clmulni_intel 
drm ata_piix intel_agp libata aesni_intel intel_gtt aes_x86_64 lrw gf128mul 
glue_helper uhci_hcd ablk_helper ehci_pci cryptd ehci_hcd agpgart scsi_mod 
virtio_pci psmouse usbcore virtio_ring virtio usb_common floppy button
Oct  8 11:04:03 ubuntu kernel: [   31.788400] CPU: 0 PID: 1807 Comm: nbd-client 
Not tainted 4.6.0-040600-generic #201606100558
Oct  8 11:04:03 ubuntu kernel: [   31.788402] Hardware name: QEMU Standard PC 
(i440FX + PIIX, 1996), BIOS 0.0.0 02/06/2015
Oct  8 11:04:03 ubuntu kernel: [   31.788404]  0086 
47e82912 88039807faa8 8135f483
Oct  8 11:04:03 ubuntu kernel: [   31.788409]   
 88039807fae8 810814db
Oct  8 11:04:03 ubuntu kernel: [   31.788412]  009647e82912 
0200 8803987d8840 880393e74070
Oct  8 11:04:03 ubuntu kernel: [   31.788416] Call Trace:
Oct  8 11:04:03 ubuntu kernel: [   31.788445]  [] 
dump_stack+0x63/0x90
Oct  8 11:04:03 ubuntu kernel: [   31.788448]  [] 
__warn+0xcb/0xf0
Oct  8 11:04:03 ubuntu kernel: [   31.788451]  [] 
warn_slowpath_null+0x1d/0x20
Oct  8 11:04:03 ubuntu kernel: [   31.788455]  [] 
__local_bh_enable_ip+0x6b/0x80
Oct  8 11:04:03 ubuntu kernel: [   31.788461]  [] 
lock_sock_nested+0x57/0x70
Oct  8 11:04:03 ubuntu kernel: [   31.788471]  [] 
inet_shutdown+0x3b/0x110
Oct  8 11:04:03 ubuntu kernel: [   31.788474]  [] 
kernel_sock_shutdown+0x10/0x20
Oct  8 11:04:03 ubuntu kernel: [   31.788481]  [] 
sock_shutdown+0x4a/0xa0 [nbd]
Oct  8 11:04:03 ubuntu kernel: [   31.788486]  [] 
__nbd_ioctl+0x615/0xb70 [nbd]
Oct  8 11:04:03 ubuntu kernel: [   31.788492]  [] ? 
select_idle_sibling+0x2a/0x120
Oct  8 11:04:03 ubuntu kernel: [   31.788497]  [] ? 
select_task_rq_fair+0x61b/0x730
Oct  8 11:04:03 ubuntu kernel: [   31.788504]  [] ? 
try_to_wake_up+0x47/0x3c0
Oct  8 11:04:03 ubuntu kernel: [   31.788507]  [] ? 
default_wake_function+0x12/0x20
Oct  8 11:04:03 ubuntu kernel: [   31.788519]  [] ? 
pollwake+0x73/0x90
Oct  8 11:04:03 ubuntu kernel: [   31.788526]  [] ? 
wake_up_q+0x70/0x70
Oct  8 11:04:03 ubuntu kernel: [   31.788532]  [] ? 
__wake_up_common+0x52/0x90
Oct  8 11:04:03 ubuntu kernel: [   31.788541]  [] ? 
current_fs_time+0x23/0x30
Oct  8 11:04:03 ubuntu kernel: [   31.788546]  [] ? 
security_capable+0x48/0x60
Oct  8 11:04:03 ubuntu kernel: [   31.788550]  [] 
nbd_ioctl+0x5d/0x7d [nbd]
Oct  8 11:04:03 ubuntu kernel: [   31.788557]  [] 
blkdev_ioctl+0x274/0x920
Oct  8 11:04:03 ubuntu kernel: [   31.788564]  [] ? 
__vfs_write+0xc9/0x120
Oct  8 11:04:03 ubuntu kernel: [   31.788574]  [] 
block_ioctl+0x3d/0x50
Oct  8 11:04:03 ubuntu kernel: [   31.788577]  [] 
do_vfs_ioctl+0xa1/0x5b0
Oct  8 11:04:03 ubuntu kernel: [   31.788582]  [] ? 
vfs_write+0x149/0x1a0
Oct  8 11:04:03 ubuntu kernel: [   31.788585]  [] 
SyS_ioctl+0x79/0x90
Oct  8 11:04:03 ubuntu kernel: [   31.788601]  [] 
entry_SYSCALL_64_fastpath+0x1e/0xa8
Oct  8 11:04:03 ubuntu kernel: [   31.788605] ---[ end trace 017ff4e61fb69ab3 
]---
Oct  8 11:04:03 ubuntu kernel: [   31.788687] block nbd0: Receive control 
failed (result -32)
Oct  8 11:04:05 ubuntu kernel: [   33.299189] nbd: unregistered device at major 
43

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1793464

Title:
  Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This issue can be reproduced on AMD64 

[Kernel-packages] [Bug 1793464] Re: Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

2018-10-08 Thread Colin Ian King
And occurs way back to 4.4.0-21 too.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1793464

Title:
  Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This issue can be reproduced on AMD64 KVM //bare-metal node, s390x
  zKVM node

  The test itself will pass, but the system will hang after a few
  second.

  Steps:
    1. Deploy a node with 4.4 Xenial
    2. Run the ubuntu_nbd_smoke_test

  If you have access to the console, you will see that this system
  actually bumped into a kernel panic:

   Unable to handle kernel pointer dereference in virtual kernel address space
   failing address: 03ff802c1000 TEID: 03ff802c1803
   Fault in home space mode while using kernel ASCE.
   Log here (s390x KVM): https://pastebin.ubuntu.com/p/dNmtvbGjmz/

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-136-generic 4.4.0-136.162
  ProcVersionSignature: Ubuntu 4.4.0-136.162-generic 4.4.144
  Uname: Linux 4.4.0-136-generic s390x
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  Date: Thu Sep 20 03:46:00 2018
  HibernationDevice: RESUME=UUID=ca468a9c-9563-442c-85c6-6055e800a66e
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig'
  Lspci:

  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:

  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=UUID=b65b756a-ba4e-4c53-aa32-0db2bdb50bb3 
crashkernel=196M
  RelatedPackageVersions:
   linux-restricted-modules-4.4.0-136-generic N/A
   linux-backports-modules-4.4.0-136-generic  N/A
   linux-firmware 1.157.20
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1793464/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1793464] Re: Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

2018-10-08 Thread Colin Ian King
Can reproduce this on a -133 xenial kernel too on Xenial VM.  After the
test, I ran vmstat 1 and observed the machine just hang after ~20 or so
seconds.

** Changed in: linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu)
   Importance: Undecided => High

** Changed in: linux (Ubuntu)
   Status: Incomplete => In Progress

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1793464

Title:
  Kernel panic after the ubuntu_nbd_smoke_test on Xenial kernel

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This issue can be reproduced on AMD64 KVM //bare-metal node, s390x
  zKVM node

  The test itself will pass, but the system will hang after a few
  second.

  Steps:
    1. Deploy a node with 4.4 Xenial
    2. Run the ubuntu_nbd_smoke_test

  If you have access to the console, you will see that this system
  actually bumped into a kernel panic:

   Unable to handle kernel pointer dereference in virtual kernel address space
   failing address: 03ff802c1000 TEID: 03ff802c1803
   Fault in home space mode while using kernel ASCE.
   Log here (s390x KVM): https://pastebin.ubuntu.com/p/dNmtvbGjmz/

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-136-generic 4.4.0-136.162
  ProcVersionSignature: Ubuntu 4.4.0-136.162-generic 4.4.144
  Uname: Linux 4.4.0-136-generic s390x
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay'
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  Date: Thu Sep 20 03:46:00 2018
  HibernationDevice: RESUME=UUID=ca468a9c-9563-442c-85c6-6055e800a66e
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig'
  Lspci:

  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:

  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=UUID=b65b756a-ba4e-4c53-aa32-0db2bdb50bb3 
crashkernel=196M
  RelatedPackageVersions:
   linux-restricted-modules-4.4.0-136-generic N/A
   linux-backports-modules-4.4.0-136-generic  N/A
   linux-firmware 1.157.20
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill'
  SourcePackage: linux
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1793464/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
** Changed in: linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu)
   Importance: Undecided => Medium

** Changed in: linux (Ubuntu)
   Status: Incomplete => In Progress

** Summary changed:

- timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency
+ sysfs test ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  sysfs test ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  In Progress

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEB

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
And reproducible way back to 4.4.0-21-lowlatency too

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  timer test in ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 DEBUG| [stdout] [ 1422.281009]  [] ? 
find_vma+0x68/0x70
  16:50:05 DEBUG| 

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
And is reproduced in previous kernel 4.4.0-137-lowlatency too, so not a
regression on -138

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  timer test in ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 DEBUG| [stdout] [ 1422.281009]  [] ? 

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
..after this hangs, I see:

[  261.473664] Oops:  [#1] PREEMPT SMP 
[  261.520488] Modules linked in: ipmi_ssif ipmi_devintf kvm_amd kvm dcdbas 
irqbypass input_leds serio_raw joydev amd64_edac_mod edac_mce_amd fam15h_power 
edac_core k10temp ipmi_si shpchp i2c_piix4 ipmi_msghandler 8250_fintek mac_hid 
acpi_power_meter ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core ib_addr 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 btrfs raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
hid_generic ghash_clmulni_intel aesni_intel mptsas aes_x86_64 pata_acpi lrw 
mptscsih gf128mul usbhid ahci glue_helper ablk_helper hid cryptd psmouse 
pata_atiixp mptbase libahci bnx2 scsi_transport_sas
[  262.283269] CPU: 3 PID: 1846 Comm: stress-ng Not tainted 
4.4.0-138-lowlatency #164-Ubuntu
[  262.381038] Hardware name: Dell Inc. PowerEdge R415/08WNM9, BIOS 1.9.3 
04/26/2012
[  262.470483] task: 8800d8323900 ti: 8800ded34000 task.ti: 
8800ded34000
[  262.559926] RIP: 0010:[]  [] 
kmem_cache_alloc+0x72/0x200
[  262.660817] RSP: 0018:8800ded37d18  EFLAGS: 00010206
[  262.724260] RAX:  RBX: 88020f630e10 RCX: 22a95403
[  262.809545] RDX: 22a95203 RSI: 22a95203 RDI: 0001a740
[  262.894829] RBP: 8800ded37d48 R08: 811d8ebc R09: 0001
[  262.980114] R10: 88021eff9000 R11:  R12: 02000200
[  263.065397] R13: 7ffcfd834000 R14: 880215003b00 R15: 880215003b00
[  263.150683] FS:  7f0599fc9700() GS:8802156c() 
knlGS:
[  263.247404] CS:  0010 DS:  ES:  CR0: 80050033
[  263.316050] CR2: 7ffcfd834000 CR3: d83a6000 CR4: 000406f0
[  263.401334] Stack:
[  263.425258]  7f059881dfff 88020f630e10  
8800d3d84f40
[  263.513690]  0001 8800dae95a40 8800ded37da0 
811d8ebc
[  263.602120]   88020ef23518 88020f630e88 
88020ef23590
[  263.690550] Call Trace:
[  263.719678]  [] anon_vma_clone+0x6c/0x200
[  263.786248]  [] anon_vma_fork+0x32/0x140
[  263.851779]  [] copy_process+0x1474/0x1c70
[  263.919385]  [] _do_fork+0x80/0x390
[  263.979714]  [] SyS_clone+0x19/0x20
[  264.040041]  [] entry_SYSCALL_64_fastpath+0x22/0xc1
[  264.117009] Code: 08 48 39 f2 75 e7 48 83 78 10 00 4c 8b 28 0f 84 40 01 00 
00 4d 85 ed 0f 84 37 01 00 00 49 63 46 20 49 8b 3e 48 8d 8a 00 02 00 00 <49> 8b 
5c 05 00 4c 89 e8 65 48 0f c7 0f 0f 94 c0 84 c0 74 ab 49 
[  264.343111] RIP  [] kmem_cache_alloc+0x72/0x200
[  264.415924]  RSP 
[  264.457529] CR2: 7ffcfd834000
[  264.497083] ---[ end trace 9af64fddce7496da ]---

and further opens seem to hang up.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  timer test in ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp 

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
It appears to hang when more than one thread is concurrently opening
/sys/kernel/security/tpm0/ascii_bios_measurements

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  timer test in ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 DEBUG| 

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
..and after that last oops we get:

ubuntu@gonzo:~$ dmesg
Killed
ubuntu@gonzo:~$ top
Killed
ubuntu@gonzo:~$ ps
Killed

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  timer test in ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 1421.187782] R10: 88000ff0 R11: 
0ff0 R12: 024000c0
  16:50:05 DEBUG| [stdout] [ 1421.273070] R13: 811d53e8 R14: 
880215003b00 R15: 880215003b00
  16:50:05 DEBUG| [stdout] [ 1421.358354] FS:  7f1323076700() 
GS:8802156c() knlGS:
  16:50:05 DEBUG| [stdout] [ 1421.455081] CS:  0010 DS:  ES:  CR0: 
80050033
  16:50:05 DEBUG| [stdout] [ 1421.523728] CR2: 00010001 CR3: 
00016f6dc000 CR4: 000406f0
  16:50:05 DEBUG| [stdout] [ 1421.609013] Stack:
  16:50:05 DEBUG| [stdout] [ 1421.632941]   88019c080a28 
 cfff
  16:50:05 DEBUG| [stdout] [ 1421.721371]  7fffefec 88020ff29000 
88016a5b3c38 811d53e8
  16:50:05 DEBUG| [stdout] [ 1421.809805]  0002 88019c080a28 
cfff 7fffefec
  16:50:05 DEBUG| [stdout] [ 1421.898237] Call Trace:
  16:50:05 DEBUG| [stdout] [ 1421.927367]  [] 
anon_vma_prepare+0x48/0x180
  16:50:05 DEBUG| [stdout] [ 1421.996017]  [] 
handle_mm_fault+0x13ed/0x1b70
  16:50:05 DEBUG| [stdout] [ 1422.066743]  [] ? 
atime_needs_update+0x6f/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.138510]  [] ? 
touch_atime+0x33/0xd0
  16:50:05 DEBUG| [stdout] [ 1422.202997]  [] ? 
generic_file_read_iter+0x5dc/0x6b0
  16:50:05 DEBUG| 

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
Exercising /sys/kernel/security/tpm0/ascii_bios_measurements causes:

[  381.545913] iounmap: bad address c9000e62
[  381.602107] CPU: 2 PID: 1833 Comm: stress-ng-sysfs Not tainted 
4.4.0-138-lowlatency #164-Ubuntu
[  381.602109] Hardware name: Dell Inc. PowerEdge R415/08WNM9, BIOS 1.9.3 
04/26/2012
[  381.602111]  0286 fc02aae09141f60c 8800daa03bc0 
8140dc61
[  381.602117]  88021145cbc0 c9000e62 8800daa03be0 
8106ec1f
[  381.602124]  88021145cbc0 c9000e62 8800daa03bf0 
8106ec5c
[  381.602127] Call Trace:
[  381.602134]  [] dump_stack+0x63/0x82
[  381.602138]  [] iounmap.part.1+0x7f/0x90
[  381.602140]  [] iounmap+0x2c/0x30
[  381.602144]  [] acpi_os_map_cleanup.part.9+0x31/0x40
[  381.602147]  [] acpi_os_unmap_iomem+0xbe/0xf0
[  381.602151]  [] read_log+0xad/0x170
[  381.602153]  [] tpm_ascii_bios_measurements_open+0x37/0x90
[  381.602158]  [] do_dentry_open+0x202/0x310
[  381.602160]  [] ? 
tpm_binary_bios_measurements_open+0x90/0x90
[  381.602164]  [] vfs_open+0x54/0x80
[  381.602166]  [] ? may_open+0x5b/0xf0
[  381.602169]  [] path_openat+0x1b6/0x13a0
[  381.602172]  [] ? free_one_page+0x191/0x340
[  381.602175]  [] do_filp_open+0x91/0x100
[  381.602179]  [] ? common_file_perm+0x70/0x1b0
[  381.602181]  [] ? __alloc_fd+0xc7/0x190
[  381.602185]  [] do_sys_open+0x138/0x2b0
[  381.602187]  [] SyS_open+0x1e/0x20
[  381.602191]  [] entry_SYSCALL_64_fastpath+0x22/0xc1

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1796250

Title:
  timer test in ubuntu_stress_smoke_test will cause kernel oops on
  X-lowlatency

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  This is found on a SRU testing node "gonzo", with 4.4 amd64 lowlatency
  kernel.

  This issue cannot be reproduced with the kernel 4.4.0-137 in -updates (a bit 
random, see comment #4)
  And cannot be reproduced with the AMD64 generic kernel in -proposed 
(4.4.0-138) as well

  16:50:01 DEBUG| [stdout] timer STARTING
  16:50:05 ERROR| [stderr] 
/home/ubuntu/autotest/client/tests/ubuntu_stress_smoke_test/ubuntu_stress_smoke_test.sh:
 line 111: 39506 Killed  ./stress-ng -v -t ${DURATION} --${s} 
${INSTANCES} ${STRESS_OPTIONS} &> ${TMP_FILE}
  16:50:05 DEBUG| [stdout] timer RETURNED 137
  16:50:05 DEBUG| [stdout] timer FAILED (kernel oopsed)
  16:50:05 DEBUG| [stdout] [ 1418.982110] BUG: unable to handle kernel paging 
request at 00010001
  16:50:05 DEBUG| [stdout] [ 1419.065329] IP: [] 
kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1419.137102] PGD 16f6dd067 PUD 0
  16:50:05 DEBUG| [stdout] [ 1419.175602] Oops:  [#6] SMP
  16:50:05 DEBUG| [stdout] [ 1419.214101] Modules linked in: unix_diag 
binfmt_misc vhost_net vhost macvtap cuse macvlan dccp_ipv4 dccp 
jitterentropy_rng algif_rng ghash_generic salsa20_generic salsa20_x86_64 
camellia_generic camellia_aesni_avx_x86_64 camellia_x86_64 cast6_avx_x86_64 
cast6_generic cast_common serpent_avx_x86_64 serpent_sse2_x86_64 
serpent_generic twofish_generic twofish_avx_x86_64 twofish_x86_64_3way 
twofish_x86_64 twofish_common xts algif_skcipher tgr192 wp512 rmd320 rmd256 
rmd160 rmd128 md4 algif_hash af_alg aufs kvm_amd kvm ipmi_devintf ipmi_ssif 
irqbypass dcdbas ipmi_si fam15h_power acpi_power_meter joydev input_leds 
ipmi_msghandler serio_raw i2c_piix4 k10temp amd64_edac_mod 8250_fintek mac_hid 
shpchp edac_mce_amd edac_core ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi
  16:50:05 DEBUG| [stdout] [ 1420.062172]  scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel mptsas mptscsih pata_acpi hid_generic aesni_intel 
aes_x86_64 mptbase lrw gf128mul glue_helper psmouse ahci ablk_helper usbhid 
cryptd pata_atiixp libahci scsi_transport_sas bnx2 hid
  16:50:05 DEBUG| [stdout] [ 1420.472200] CPU: 3 PID: 39506 Comm: 
ubuntu_stress_s Tainted: G  D 4.4.0-138-generic #164-Ubuntu
  16:50:05 DEBUG| [stdout] [ 1420.588693] Hardware name: Dell Inc. PowerEdge 
R415/08WNM9, BIOS 1.9.3 04/26/2012
  16:50:05 DEBUG| [stdout] [ 1420.678138] task: 880177823800 ti: 
88016a5b task.ti: 88016a5b
  16:50:05 DEBUG| [stdout] [ 1420.767584] RIP: 0010:[]  
[] kmem_cache_alloc+0x77/0x1f0
  16:50:05 DEBUG| [stdout] [ 1420.868478] RSP: 0018:88016a5b3bd0  EFLAGS: 
00010202
  16:50:05 DEBUG| [stdout] [ 1420.931924] RAX:  RBX: 
024000c0 RCX: 0129e216
  16:50:05 DEBUG| [stdout] [ 1421.017209] RDX: 0129e215 RSI: 
024000c0 RDI: 0001a5c0
  16:50:05 DEBUG| [stdout] [ 1421.102496] RBP: 88016a5b3c00 R08: 
8802156da5c0 R09: 00010001
  16:50:05 DEBUG| [stdout] [ 

[Kernel-packages] [Bug 1796250] Re: timer test in ubuntu_stress_smoke_test will cause kernel oops on X-lowlatency

2018-10-05 Thread Colin Ian King
So I believe this is the sysfs test that causes the issues, the
subsequent tee timer stressors just appear to be broken because the
sysfs stressor got the kernel into a mangled broken state.

I was able to run the tee and timer stressors on a cleanly booted
4.4.0-137-lowlatency kernel without any issues.  Running the sysfs
stressor caused a load of oopses as follows:

  740.873265] iounmap: bad address c90004da
[  740.929503] CPU: 4 PID: 1788 Comm: stress-ng-sysfs Not tainted 
4.4.0-138-lowlatency #164-Ubuntu
[  740.929509] Hardware name: Dell Inc. PowerEdge R415/08WNM9, BIOS 1.9.3 
04/26/2012
[  740.929512]  0286 5dc48b8f33ce7222 8800da813bc0 
8140dc61
[  740.929536]  8800df1f9d80 c90004da 8800da813be0 
8106ec1f
[  740.929541]  8800df1f9d80 c90004da 8800da813bf0 
8106ec5c
[  740.929544] Call Trace:
[  740.929553]  [] dump_stack+0x63/0x82
[  740.929558]  [] iounmap.part.1+0x7f/0x90
[  740.929563]  [] iounmap+0x2c/0x30
[  740.929572]  [] acpi_os_map_cleanup.part.9+0x31/0x40
[  740.929587]  [] acpi_os_unmap_iomem+0xbe/0xf0
[  740.929607]  [] read_log+0xad/0x170
[  740.929619]  [] tpm_binary_bios_measurements_open+0x37/0x90
[  740.929626]  [] do_dentry_open+0x202/0x310
[  740.929632]  [] ? 
tpm_ascii_bios_measurements_show+0x260/0x260
[  740.929645]  [] vfs_open+0x54/0x80
[  740.929656]  [] ? may_open+0x5b/0xf0
[  740.929662]  [] path_openat+0x1b6/0x13a0
[  740.929670]  [] ? __do_page_fault+0x23a/0x440
[  740.929678]  [] ? free_one_page+0x191/0x340
[  740.929686]  [] do_filp_open+0x91/0x100
[  740.929695]  [] ? mntput_no_expire+0x2e/0x1b0
[  740.929702]  [] ? __alloc_fd+0xc7/0x190
[  740.929720]  [] do_sys_open+0x138/0x2b0
[  740.929735]  [] SyS_open+0x1e/0x20
[  740.929747]  [] entry_SYSCALL_64_fastpath+0x22/0xc1
[  749.735460] BUG: unable to handle kernel paging request at 00010001
[  749.818685] IP: [] kmem_cache_alloc_trace+0x76/0x210
[  749.896706] PGD d3e68067 PUD 0 
[  749.934172] Oops:  [#1] PREEMPT SMP 
[  749.981003] Modules linked in: kvm_amd kvm irqbypass ipmi_devintf ipmi_ssif 
dcdbas input_leds joydev amd64_edac_mod 8250_fintek k10temp fam15h_power 
serio_raw edac_mce_amd ipmi_si i2c_piix4 acpi_power_meter edac_core 
ipmi_msghandler shpchp mac_hid ib_iser rdma_cm iw_cm ib_cm ib_sa ib_mad ib_core 
ib_addr iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 btrfs 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 multipath linear crct10dif_pclmul crc32_pclmul 
ghash_clmulni_intel aesni_intel hid_generic aes_x86_64 mptsas lrw mptscsih 
gf128mul usbhid glue_helper ablk_helper pata_acpi hid cryptd psmouse 
pata_atiixp mptbase ahci bnx2 libahci scsi_transport_sas
[  750.743813] CPU: 4 PID: 1791 Comm: stress-ng-sysfs Not tainted 
4.4.0-138-lowlatency #164-Ubuntu
[  750.847825] Hardware name: Dell Inc. PowerEdge R415/08WNM9, BIOS 1.9.3 
04/26/2012
[  750.937282] task: 8800d3e45580 ti: 8800da824000 task.ti: 
8800da824000
[  751.026734] RIP: 0010:[]  [] 
kmem_cache_alloc_trace+0x76/0x210
[  751.133881] RSP: 0018:8800da827c18  EFLAGS: 00010202
[  751.197334] RAX:  RBX:  RCX: 0b1af204
[  751.282627] RDX: 0b1af004 RSI: 0b1af004 RDI: 0001a740
[  751.367921] RBP: 8800da827c58 R08: 880215003b00 R09: 8129e635
[  751.453209] R10: 0637760b R11:  R12: 024000c0
[  751.538502] R13: 00010001 R14: 88020fc24900 R15: 880215003b00
[  751.623798] FS:  7f8e5a1f4700() GS:88021570() 
knlGS:
[  751.720535] CS:  0010 DS:  ES:  CR0: 80050033
[  751.789183] CR2: 00010001 CR3: db29e000 CR4: 000406f0
[  751.874477] Stack:
[  751.898415]   88020fc24900 0030 

[  751.986848]  88020f67cf00 8800d94cf980 88020fc24900 
8802124386a8
[  752.075289]  8800da827c98 8129e635  
88020fc24900
[  752.163722] Call Trace:
[  752.192853]  [] kernfs_fop_open+0x275/0x370
[  752.261505]  [] do_dentry_open+0x202/0x310
[  752.329117]  [] ? kernfs_fop_release+0x60/0x60
[  752.400893]  [] vfs_open+0x54/0x80
[  752.460186]  [] ? may_open+0x5b/0xf0
[  752.521559]  [] path_openat+0x1b6/0x13a0
[  752.587093]  [] ? bad_area_access_error+0x43/0x50
[  752.661986]  [] do_filp_open+0x91/0x100
[  752.726486]  [] ? __alloc_fd+0xc7/0x190
[  752.790980]  [] do_sys_open+0x138/0x2b0
[  752.855479]  [] SyS_open+0x1e/0x20
[  752.914775]  [] entry_SYSCALL_64_fastpath+0x22/0xc1
[  752.991744] Code: 08 48 39 f2 75 e7 48 83 78 10 00 4c 8b 28 0f 84 48 01 00 
00 4d 85 ed 0f 84 3f 01 00 00 49 63 47 20 49 8b 3f 48 8d 8a 00 02 00 00 <49> 8b 
5c 05 00 4c 89 e8 65 48 0f c7 0f 0f 94 c0 84 c0 74 ab 49 
[  753.217863] RIP  [] kmem_cache_alloc_trace+0x76/0x210
[  753.296928]  RSP 
[  753.338534] CR2: 00010001
[  

[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-10-01 Thread Colin Ian King
@Prakash, you will require the 4.15 Bionic kernel as well, this commit
has only just been committed to the kernel source, so please wait until
this is released in a couple of weeks.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in linux package in Ubuntu:
  Fix Committed
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Incomplete
Status in zfs-linux source package in Xenial:
  New
Status in linux source package in Bionic:
  Fix Committed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Committed
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, Xenial, Bionic ==

  Bug:

  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

  == Fix ==

  Upstream zfs commit 4ceb8dd6fdfd ("Fix 'zpool create -t '")

  == Test ==

  Without the fix, creating a pool with -t will show an error message
  such as:

  "cannot create 'rpool': no such pool or dataset" and will return an
  error number even though it successfully created the pool. With the
  fix, the error is not shown and it exits with 0.

  E.g.

  pool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1 
  cannot create 'rpool': no such pool or dataset

  == Regression Potential ==

  Small. This is an upstream fix that is know to fix the error. There
  are no known subsequent changes to the fix hence suggesting it is a
  good fix. The change is very limited and testing shows it does not
  break any of the kernel team's ZFS regression testing.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-09-27 Thread Colin Ian King
Tested 0.6.5.6-0ubuntu25 and it works without any issues, so marking
this as verified.

** Tags removed: verification-required-xenial
** Tags added: verification-done-xenial

** Changed in: linux (Ubuntu Xenial)
 Assignee: tenox (senseimyijaki) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu Bionic)
 Assignee: tenox (senseimyijaki) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu Bionic)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu Xenial)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu Bionic)
   Importance: Undecided => High

** Changed in: zfs-linux (Ubuntu Xenial)
   Importance: Undecided => High

** Changed in: linux (Ubuntu Xenial)
   Importance: Undecided => High

** Changed in: linux (Ubuntu Bionic)
   Importance: Undecided => High

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Released
Status in zfs-linux source package in Xenial:
  Fix Committed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing b

[Kernel-packages] [Bug 1782565] Re: s390 illegal operation

2018-09-17 Thread Colin Ian King
** Changed in: linux (Ubuntu Bionic)
   Status: Expired => New

** Changed in: linux (Ubuntu Cosmic)
   Status: Expired => New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1782565

Title:
  s390 illegal operation

Status in linux package in Ubuntu:
  Incomplete
Status in linux source package in Bionic:
  Incomplete
Status in linux source package in Cosmic:
  Incomplete

Bug description:
  Managed to break sysfs when running:

  sudo ./stress-ng  --sysfs 4 -t 60 -v --verify

  
  Kernel: 4.15.0-29-generic and 4.17 too, this seems to occur on previous 4.15 
kernels but I've not tested too far back.

  
  [   45.057702] [ cut here ]
  [   45.057704] kernel BUG at 
/build/linux-4OqUyO/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [   45.057773] illegal operation: 0001 ilc:1 [#1] SMP 
  [   45.057775] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi nfsd auth_rpcgss nfs_acl lockd grace 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear crc32_vx_s390 virtio_net virtio_blk
  [   45.057805] CPU: 0 PID: 5056 Comm: stress-ng-sysfs Tainted: P   O  
   4.15.0-29-generic #31-Ubuntu
  [   45.057806] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [   45.057809] Krnl PSW : 2a20609d 800ebb38 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [   45.057814]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [   45.057816] Krnl GPRS: de6dc5c2778317a7 7ff93a20 0080 
1164
  [   45.057818]03ff800058da 1165 6b9cc840 
7061a100
  [   45.057819] 0068feb8 72c30d80 
71972000
  [   45.057821]03ff80008210 1000 03ff800058da 
70a47ce8
  [   45.057830] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
03ff80005918: c0f40560  brcl
15,3ff800063d8
   #03ff8000591e: a7f40001  brc 
15,3ff80005920
   >03ff80005922: 0707  bcr 0,%r7
03ff80005924: 0707  bcr 0,%r7
03ff80005926: 0707  bcr 0,%r7
03ff80005928: c004  brcl
0,3ff80005928
03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [   45.057844] Call Trace:
  [   45.057847] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
  [   45.057852]  [<0068fef4>] dev_attr_show+0x3c/0x80 
  [   45.057858]  [<00424414>] sysfs_kf_seq_show+0xbc/0x1a8 
  [   45.057862]  [<003afddc>] seq_read+0xec/0x4c8 
  [   45.057865]  [<00381bba>] vfs_read+0x8a/0x150 
  [   45.057867]  [<00382156>] SyS_read+0x66/0xe0 
  [   45.057872]  [<008e2350>] system_call+0xdc/0x2c8 
  [   45.057873] Last Breaking-Event-Address:
  [   45.057875]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
  [   45.057876]  
  [   45.057877] ---[ end trace 49fd8401e53e0e99 ]---

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1782565/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1792973] Re: ubuntu_nbd_smoke_test failed on 4.15 Trusty Azure

2018-09-17 Thread Colin Ian King
I've pushed a potential set of fixes for this test:
http://kernel.ubuntu.com/git/ubuntu/autotest-client-tests.git/commit/?h
=master-next=ed938798b02edda1dd639edc754e62ea4a1db2d3

I ran this fix successfully many times for over 1h on a i386 VM and it
seemed to fix similar issues I was seeing, so I think it may help.

Can this be tested to see if it fixes the issue you are seeing?

** Changed in: ubuntu-kernel-tests
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: ubuntu-kernel-tests
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-signed-azure in Ubuntu.
https://bugs.launchpad.net/bugs/1792973

Title:
  ubuntu_nbd_smoke_test failed on 4.15 Trusty Azure

Status in ubuntu-kernel-tests:
  New
Status in linux-signed-azure package in Ubuntu:
  New

Bug description:
  The ubuntu_nbd_smoke test will timeout on a 4.15 Trusty Azure node.

  The test will get killed by the timeout setting and the job hangs.

  There is also a rmmod related error message:
   [stderr] rmmod: ERROR: Module nbd is not currently loaded


   Running 'DEBIAN_FRONTEND=noninteractive apt-get install --yes --force-yes 
nbd-server nbd-client gdb'
   [stdout] Reading package lists...
   [stdout] Building dependency tree...
   [stdout] Reading state information...
   [stdout] gdb is already the newest version.
   [stdout] The following NEW packages will be installed:
   [stdout]   nbd-client nbd-server
   [stdout] 0 upgraded, 2 newly installed, 0 to remove and 7 not upgraded.
   [stdout] Need to get 104 kB of archives.
   [stdout] After this operation, 430 kB of additional disk space will be used.
   [stdout] Get:1 http://azure.archive.ubuntu.com/ubuntu/ trusty-updates/main 
nbd-client amd64 1:3.7-1ubuntu0.1 [45.1 kB]
   [stdout] Get:2 http://azure.archive.ubuntu.com/ubuntu/ trusty-updates/main 
nbd-server amd64 1:3.7-1ubuntu0.1 [58.8 kB]
   [stdout] Preconfiguring packages ...
   [stdout] Fetched 104 kB in 0s (9,346 kB/s)
   [stdout] Selecting previously unselected package nbd-client.
   [stdout] (Reading database ... 54762 files and directories currently 
installed.)
   [stdout] Preparing to unpack .../nbd-client_1%3a3.7-1ubuntu0.1_amd64.deb ...
   [stdout] Unpacking nbd-client (1:3.7-1ubuntu0.1) ...
   [stdout] Selecting previously unselected package nbd-server.
   [stdout] Preparing to unpack .../nbd-server_1%3a3.7-1ubuntu0.1_amd64.deb ...
   [stdout] Unpacking nbd-server (1:3.7-1ubuntu0.1) ...
   [stdout] Processing triggers for man-db (2.6.7.1-1ubuntu1) ...
   [stdout] Processing triggers for ureadahead (0.100.0-16) ...
   [stdout] Setting up nbd-client (1:3.7-1ubuntu0.1) ...
   [stderr] Stopping NBD client process: 
   [stderr] rmmod: ERROR: Module nbd is not currently loaded
   [stderr] nbd-client.
   [stderr] Starting NBD client process: Connecting...Activating...
   [stderr] nbd-client.
   [stderr] update-initramfs: deferring update (trigger activated)
   [stdout] Setting up nbd-server (1:3.7-1ubuntu0.1) ...
   [stderr] 
   [stderr] Creating config file /etc/nbd-server/config with new version
   [stderr] Adding system user `nbd' (UID 108) ...
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] Adding new group `nbd' (GID 113) ...
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] Adding new user `nbd' (UID 108) with group `nbd' ...
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] sent invalidate(passwd) request, exiting
   [stderr] sent invalidate(group) request, exiting
   [stderr] Not creating home directory `/etc/nbd-server'.
   [stderr] 
   [stderr] ** (process:2877): WARNING **: Could not parse config file: The 
config file does not specify any exports
   [stderr] ** Message: No configured exports; quitting.
   [stderr]  nbd-server.
   [stdout] Processing triggers for ureadahead (0.100.0-16) ...
   [stdout] Processing triggers for initramfs-tools (0.103ubuntu4.11) ...
   [stdout] update-initramfs: Generating /boot/initrd.img-4.15.0-1023-azure
   true'
   Running 
'/home/azure/autotest/client/tests/ubuntu_nbd_smoke_test/ubuntu_nbd_smoke_test.sh'
   [stdout]  
   [stdout] 

   [stdout] Image path:/tmp/nbd_ima

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-09-14 Thread Colin Ian King
Diff: http://launchpadlibrarian.net/388404653/zfs-
linux_0.6.5.6-0ubuntu24_0.6.5.6-0ubuntu25.diff.gz

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Released
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 1 >/sys/module/spl/parameters/spl_taskq_kick

  but did not manage to continue.
  

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-09-14 Thread Colin Ian King
Wrong URL, ignore that.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Released
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 1 >/sys/module/spl/parameters/spl_taskq_kick

  but did not manage to continue.
  Include any warning/errors/backtraces from the system logs
  dmesg output

  [  

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-09-14 Thread Colin Ian King
Diff:

http://launchpadlibrarian.net/388038683/zfs-
linux_0.6.5.6-0ubuntu24_0.6.5.6-0ubuntu25.diff.gz

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Released
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 1 >/sys/module/spl/parameters/spl_taskq_kick

  but did not manage to continue.
  

[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-09-14 Thread Colin Ian King
Yes, it's being SRU'd - changes have to land in the kernel and get
released there, so it will take a few weeks.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in linux package in Ubuntu:
  Fix Committed
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Incomplete
Status in zfs-linux source package in Xenial:
  New
Status in linux source package in Bionic:
  Incomplete
Status in zfs-linux source package in Bionic:
  New
Status in linux source package in Cosmic:
  Fix Committed
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, Xenial, Bionic ==

  Bug:

  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

  == Fix ==

  Upstream zfs commit 4ceb8dd6fdfd ("Fix 'zpool create -t '")

  == Test ==

  Without the fix, creating a pool with -t will show an error message
  such as:

  "cannot create 'rpool': no such pool or dataset" and will return an
  error number even though it successfully created the pool. With the
  fix, the error is not shown and it exits with 0.

  E.g.

  pool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1 
  cannot create 'rpool': no such pool or dataset

  == Regression Potential ==

  Small. This is an upstream fix that is know to fix the error. There
  are no known subsequent changes to the fix hence suggesting it is a
  good fix. The change is very limited and testing shows it does not
  break any of the kernel team's ZFS regression testing.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-09-12 Thread Colin Ian King
** Description changed:

+ == SRU Justification, Xenial, Bionic ==
+ 
+ Bug:
+ 
  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears to
  be successfully created.
  
  For example:
  
  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -
  
  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502
  
  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509
  
  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?
  
  Thanks,
  Prakash
+ 
+ == Fix ==
+ 
+ Upstream zfs commit 4ceb8dd6fdfd ("Fix 'zpool create -t '")
+ 
+ == Test ==
+ 
+ Without the fix, creating a pool with -t will show an error message such
+ as:
+ 
+ "cannot create 'rpool': no such pool or dataset" and will return an
+ error number even though it successfully created the pool. With the fix,
+ the error is not shown and it exits with 0.
+ 
+ E.g.
+ 
+ pool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1 
+ cannot create 'rpool': no such pool or dataset
+ 
+ == Regression Potential ==
+ 
+ Small. This is an upstream fix that is know to fix the error. There are
+ no known subsequent changes to the fix hence suggesting it is a good
+ fix. The change is very limited and testing shows it does not break any
+ of the kernel team's ZFS regression testing.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in linux package in Ubuntu:
  Incomplete
Status in zfs-linux package in Ubuntu:
  In Progress
Status in linux source package in Xenial:
  Incomplete
Status in zfs-linux source package in Xenial:
  New
Status in linux source package in Bionic:
  Incomplete
Status in zfs-linux source package in Bionic:
  New
Status in linux source package in Cosmic:
  Incomplete
Status in zfs-linux source package in Cosmic:
  In Progress

Bug description:
  == SRU Justification, Xenial, Bionic ==

  Bug:

  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

  == Fix ==

  Upstream zfs commit 4ceb8dd6fdfd ("Fix 'zpool create -t '")

  == Test ==

  Without the fix, creating a pool with -t will show an error message
  such as:

  "cannot create 'rpool': no such pool or dataset" and will return an
  error number even though it successfully created the pool. With the
  fix, the error is not shown and it exits with 0.

  E.g.

  pool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1 
  cannot create 'rpool': no such pool or dataset

  == Regression Potential ==

  Small. This is an upstream fix that is know to fix the error. There
  are no known subsequent changes to the fix hence suggesting it is a
  good fix. The change is very limited and testing shows it does not
  break any of the kernel team's ZFS regression testing.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-09-12 Thread Colin Ian King
** Also affects: zfs-linux (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Also affects: zfs-linux (Ubuntu Cosmic)
   Importance: Medium
 Assignee: Colin Ian King (colin-king)
   Status: In Progress

** Also affects: zfs-linux (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Also affects: linux (Ubuntu)
   Importance: Undecided
   Status: New

** Changed in: linux (Ubuntu Cosmic)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu Bionic)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu Xenial)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu Bionic)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu Xenial)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu Bionic)
   Importance: Undecided => Medium

** Changed in: zfs-linux (Ubuntu Xenial)
   Importance: Undecided => Medium

** Changed in: linux (Ubuntu Cosmic)
   Importance: Undecided => Medium

** Changed in: linux (Ubuntu Bionic)
   Importance: Undecided => Medium

** Changed in: linux (Ubuntu Xenial)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in linux package in Ubuntu:
  New
Status in zfs-linux package in Ubuntu:
  In Progress
Status in linux source package in Xenial:
  New
Status in zfs-linux source package in Xenial:
  New
Status in linux source package in Bionic:
  New
Status in zfs-linux source package in Bionic:
  New
Status in linux source package in Cosmic:
  New
Status in zfs-linux source package in Cosmic:
  In Progress

Bug description:
  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-09-12 Thread Colin Ian King
** Tags removed: verification-done-xenial
** Tags added: verification-required-xenial

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Released
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 1 >/sys/module/spl/parameters/spl_taskq_kick

  but did not manage to continue.
  Include any 

[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-09-12 Thread Colin Ian King
Thanks for the fix. I will add this to Cosmic and SRU this for Bionic
and Xenial.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1654517] Re: ZFS I/O hangs for minutes

2018-09-12 Thread Colin Ian King
This bug has been waiting for feedback for questions in comment #35 and
#36 for months. I'm going to mark this bug as Won't fix because of
limited responses. If this bug is still an issue, please re-open it.

** Changed in: zfs-linux (Ubuntu)
   Status: Incomplete => Won't Fix

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1654517

Title:
  ZFS I/O hangs for minutes

Status in Native ZFS for Linux:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Won't Fix

Bug description:
  I/O for multiple programs, like `thunderbird`, `firefox`, etc., hangs
  for minutes and approx. 100 `z_rd_int_[n]` and `z_wr_int_[n]` kernel
  threads are created, `dmesg` contains

  [ 9184.451606] INFO: task txg_sync:11471 blocked for more than 120 
seconds.
  [ 9184.451610]   Tainted: P   OE   4.8.0-32-generic #34-Ubuntu
  [ 9184.451611] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" 
disables this message.
  [ 9184.451612] txg_syncD a240ab3a7aa8 0 11471  2 
0x
  [ 9184.451616]  a240ab3a7aa8 00ffbb6ade1f a24095148000 
a240e5ca5580
  [ 9184.451618]  0046 a240ab3a8000 a240ff359200 
7fff
  [ 9184.451620]  a23d36cf9050 0001 a240ab3a7ac0 
bbe96b15
  [ 9184.451621] Call Trace:
  [ 9184.451627]  [] schedule+0x35/0x80
  [ 9184.451628]  [] schedule_timeout+0x22a/0x3f0
  [ 9184.451631]  [] ? __switch_to+0x2ce/0x6c0
  [ 9184.451633]  [] ? pick_next_task_fair+0x48c/0x4c0
  [ 9184.451635]  [] ? ktime_get+0x41/0xb0
  [ 9184.451636]  [] io_schedule_timeout+0xa4/0x110
  [ 9184.451644]  [] cv_wait_common+0xb2/0x130 [spl]
  [ 9184.451646]  [] ? wake_atomic_t_function+0x60/0x60
  [ 9184.451650]  [] __cv_wait_io+0x18/0x20 [spl]
  [ 9184.451689]  [] zio_wait+0xfd/0x1d0 [zfs]
  [ 9184.451716]  [] dsl_pool_sync+0xb8/0x480 [zfs]
  [ 9184.451745]  [] spa_sync+0x37f/0xb30 [zfs]
  [ 9184.451747]  [] ? default_wake_function+0x12/0x20
  [ 9184.451779]  [] txg_sync_thread+0x3a5/0x600 [zfs]
  [ 9184.451807]  [] ? txg_delay+0x160/0x160 [zfs]
  [ 9184.451811]  [] thread_generic_wrapper+0x71/0x80 
[spl]
  [ 9184.451815]  [] ? __thread_exit+0x20/0x20 [spl]
  [ 9184.451817]  [] kthread+0xd8/0xf0
  [ 9184.451819]  [] ret_from_fork+0x1f/0x40
  [ 9184.451821]  [] ? kthread_create_on_node+0x1e0/0x1e0
  [ 9184.451849] INFO: task mozStorage #2:21607 blocked for more than 120 
seconds.
  [ 9184.451851]   Tainted: P   OE   4.8.0-32-generic #34-Ubuntu
  [ 9184.451852] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" 
disables this message.
  [ 9184.451853] mozStorage #2   D a23fe8a5bd38 0 21607  19750 
0x0004
  [ 9184.451855]  a23fe8a5bd38 00ffa240ee8feb40 a240ecf72ac0 
a2403803b900
  [ 9184.451857]  bc2c02f7 a23fe8a5c000 a240aa940828 
a240aa940800
  [ 9184.451858]  a240aa940980  a23fe8a5bd50 
bbe96b15
  [ 9184.451860] Call Trace:
  [ 9184.451861]  [] schedule+0x35/0x80
  [ 9184.451866]  [] cv_wait_common+0x110/0x130 [spl]
  [ 9184.451868]  [] ? wake_atomic_t_function+0x60/0x60
  [ 9184.451872]  [] __cv_wait+0x15/0x20 [spl]
  [ 9184.451904]  [] zil_commit.part.11+0x79/0x7a0 [zfs]
  [ 9184.451909]  [] ? tsd_hash_search.isra.0+0x46/0xa0 
[spl]
  [ 9184.451913]  [] ? tsd_set+0x2b4/0x500 [spl]
  [ 9184.451914]  [] ? mutex_lock+0x12/0x30
  [ 9184.451945]  [] zil_commit+0x17/0x20 [zfs]
  [ 9184.451975]  [] zfs_fsync+0x7a/0xf0 [zfs]
  [ 9184.452005]  [] zpl_fsync+0x68/0xa0 [zfs]
  [ 9184.452008]  [] vfs_fsync_range+0x4b/0xb0
  [ 9184.452010]  [] do_fsync+0x3d/0x70
  [ 9184.452011]  [] SyS_fsync+0x10/0x20
  [ 9184.452013]  [] entry_SYSCALL_64_fastpath+0x1e/0xa8
  [ 9184.452023] INFO: task bitcoin-msghand:663 blocked for more than 120 
seconds.
  [ 9184.452024]   Tainted: P   OE   4.8.0-32-generic #34-Ubuntu
  [ 9184.452025] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" 
disables this message.
  [ 9184.452026] bitcoin-msghand D a23eeb23bd38 0   663  26994 
0x
  [ 9184.452028]  a23eeb23bd38 00ffa23eab434000 a240ecf7 
a24095148000
  [ 9184.452030]  a23eeb23bd20 a23eeb23c000 a240aa940828 
a240aa940800
  [ 9184.452031]  a240aa940980  a23eeb23bd50 
bbe96b15
  [ 9184.452033] Call Trace:
  [ 9184.452034]  [] schedule+0x35/0x80
  [ 9184.452039]  [] cv_wait_common+0x110/0x130 [spl]
  [ 9184.452041]  [] ? wake_atomic_t_function+0x60/0x60
  [ 9184.452044]  [] __cv_wait+0x15/0x20 [spl]
  [ 9184.452074]  [] zil_commit.part.11+0x79/0x7a0 [zfs]
  [ 9184.452079]  [] ? tsd_hash_search.isra.0+0x46/0xa0 
[spl]
  [ 

[Kernel-packages] [Bug 1774569] Re: gsmartcontrol, hdparm, and ZFS all refuse to talk to an apparently working Seagate Backup+ Hub drive after upgrade to 18.04

2018-09-10 Thread Colin Ian King
Since this is a physical through to block layer issue and not ZFS per-
se, I'm going to remove ZFS off this bug report.

** No longer affects: zfs-linux (Ubuntu)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1774569

Title:
  gsmartcontrol, hdparm, and ZFS all refuse to talk to an apparently
  working Seagate Backup+ Hub drive after upgrade to 18.04

Status in gsmartcontrol package in Ubuntu:
  Confirmed
Status in hdparm package in Ubuntu:
  Confirmed

Bug description:
  I recently upgraded from 17.10 to 18.04. After the upgrade, I noticed
  that my Seagate Backup+ Hub external drive was displaying a series of
  puzzling symptoms:

  1. gsmartcontrol can't get SMART data from the drive. I am pretty sure
  it used to report SMART data? Here's a log of it not working:

[hz] Warning: exit: Command line did not parse.
[app] execute_smartctl(): Smartctl binary did not execute cleanly.
[app] StorageDevice::execute_device_smartctl(): Smartctl binary did 
not execute cleanly.
[app] SmartctlParser::parse_section_info_property(): Unknown property 
"Physical block size"
[app] SmartctlParser::parse_section_info_property(): Unknown property 
"Logical Unit id"
[app] SmartctlParser::parse_section_info_property(): Unknown property 
"Temperature Warning"
[app] SmartctlParser::parse_section_data(): Unknown Data subsection 
encountered.
[hz] Warning: exit: Some SMART command to the disk failed, or there 
was a checksum error in a SMART data structure
[app] SmartctlParser::parse_section_info_property(): Unknown property 
"Physical block size"
[app] SmartctlParser::parse_section_info_property(): Unknown property 
"Logical Unit id"
[app] SmartctlParser::parse_section_info_property(): Unknown property 
"Temperature Warning"
[app] SmartctlParser::parse_section_data(): Unknown Data subsection 
encountered.

  2. hdparm used to be able to spin down the drive. I had it configured
  to spin it down after a few minutes of inactivity, in the hdparm
  config file. Now that no longer happens, and hdparm can't seem to talk
  to the drive meaningfully at all:

  [anovak@octagon ~]$ sudo hdparm -I /dev/sdb

  /dev/sdb:
  SG_IO: bad/missing sense data, sb[]:  70 00 05 00 00 00 00 0a 00 00 00 00 24 
00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00

  ATA device, with non-removable media
  Standards:
Likely used: 1
  Configuration:
Logical max current
cylinders   0   0
heads   0   0
sectors/track   0   0
--
Logical/Physical Sector size:   512 bytes
device size with M = 1024*1024:   0 MBytes
device size with M = 1000*1000:   0 MBytes 
cache/buffer size  = unknown
  Capabilities:
IORDY not likely
Cannot perform double-word IO
R/W multiple sector transfer: not supported
DMA: not supported
PIO: pio0 
  [anovak@octagon ~]$ sudo hdparm -y /dev/sdb

  /dev/sdb:
   issuing standby command
  SG_IO: bad/missing sense data, sb[]:  70 00 05 00 00 00 00 0a 00 00 00 00 24 
00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00

  I think this may be related to https://askubuntu.com/questions/1037997
  /upgraded-to-18-04-usb-harddrive-doesn-t-idle-anymore which is someone
  else having the same problem.

  3. The ZFS tools think the drive is hosed:

  [anovak@octagon ~]$ sudo zpool status hub
pool: hub
   state: UNAVAIL
  status: One or more devices could not be used because the label is missing 
or invalid.  There are insufficient replicas for the pool to continue
functioning.
  action: Destroy and re-create the pool from
a backup source.
 see: http://zfsonlinux.org/msg/ZFS-8000-5E
scan: none requested
  config:

NAME   STATE READ WRITE CKSUM
hubUNAVAIL  0 0 0  
insufficient replicas
  ata-ST6000DM003-2CY186_ZF200PC8  UNAVAIL  0 0 0

  This may be related to the drive having adopted a new /dev/disk/by-id
  name during the upgrade? I think it was "ata-
  ST6000DM003-2CY186_ZF200PC8" when I added it to my zpool by its
  symlink under /dev/disks/by-id, but now it is "usb-
  Seagate_Backup+_Hub_BK_NA8TQC87-0:0":

  [anovak@octagon ~]$ ls -lah 
/dev/disk/by-id/usb-Seagate_Backup+_Hub_BK_NA8TQC87-0\:0
  lrwxrwxrwx 1 root root 9 May 31 20:52 
/dev/disk/by-id/usb-Seagate_Backup+_Hub_BK_NA8TQC87-0:0 -> ../../sdb

  This *shouldn't* cause trouble; you should be able to export the zpool
  and re-import it under the new name. But zpool import shows nothing to
  import:

  [anovak@octagon ~]$ sudo zpool import
  no pools available to import

  And I also can't export or even destroy the busted zpool, because
  zpool doesn't think it exists for exporting or destroying purposes:


[Kernel-packages] [Bug 1733230] Re: 'zfs recv' hangs when receiving from a FreeBSD zfs

2018-09-10 Thread Colin Ian King
Ubuntu 17.10 is now end-of-life.  However, this is fixed in Ubuntu 18.04
with ZFS 0.7.5, so please consider using this now.

** Changed in: zfs-linux (Ubuntu)
   Status: Incomplete => Fix Released

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1733230

Title:
  'zfs recv' hangs when receiving from a FreeBSD zfs

Status in zfs-linux package in Ubuntu:
  Fix Released

Bug description:
  When creating a pool snapshot on a freenas/freebsd like this:

  ```
  root@freenas:~ # uname -a
  FreeBSD freenas.local 11.0-STABLE FreeBSD 11.0-STABLE #0 
r321665+25fe8ba8d06(freenas/11.0-stable): Mon Sep 25 06:24:11 UTC 2017 
root@gauntlet:/freenas-11-releng/freenas/_BE/objs/freenas-11-releng/freenas/_BE/os/sys/FreeNAS.amd64
  amd64
  freenas# zpool create tank gptid/62a26be7-cd4a-11e7-927b-0800278dcda9
  freenas# dd if=/dev/zero of=/tank/ur.bin bs=64K count=1
  freenas# zfs snap tank@example
  ```

  And trying to `zfs send | zfs recv` it to an ubuntu zfs like this:
  ```
  space@ubuntu:~$ apt-cache policy zfsutils-linux
  zfsutils-linux:
Installed: 0.6.5.11-1ubuntu3
Candidate: 0.6.5.11-1ubuntu3
Version table:
   *** 0.6.5.11-1ubuntu3 500
  500 http://us.archive.ubuntu.com/ubuntu artful/main amd64 Packages
  100 /var/lib/dpkg/status
  space@ubuntu:~$ lsb_release -rd
  Description:  Ubuntu 17.10
  Release:  17.10

  root@ubuntu:~$ ssh root@freenas zfs send tank@example > freenas11.zfssend
  root@ubuntu:~# cat ./freenas11.zfssend | zfs recv tank/something
  ```

  The `zfs recv` process burns 100% CPU and hangs forever. I have
  attached an example freenas11.zfssend file, though it's easily
  reproduced.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.10
  Package: zfsutils-linux 0.6.5.11-1ubuntu3
  ProcVersionSignature: Ubuntu 4.13.0-16.19-generic 4.13.4
  Uname: Linux 4.13.0-16-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.7-0ubuntu3
  Architecture: amd64
  Date: Sun Nov 19 22:10:15 2017
  ExecutablePath: /sbin/zfs
  InstallationDate: Installed on 2017-11-17 (2 days ago)
  InstallationMedia: Ubuntu-Server 17.10 "Artful Aardvark" - Release amd64 
(20171017.1)
  ProcEnviron:
   LANG=en_US.UTF-8
   SHELL=/bin/bash
   TERM=xterm-256color
   PATH=(custom, no user)
  SourcePackage: zfs-linux
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1733230/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1788432] Re: 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-31 Thread Colin Ian King
OK, thanks for the info so far, apologies for being slow to get back to
you.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1788432

Title:
  4.15 s390x kernel BUG at /build/linux-
  Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

Status in Ubuntu on IBM z Systems:
  Triaged
Status in linux package in Ubuntu:
  Triaged
Status in linux source package in Bionic:
  Triaged

Bug description:
  uname -a
  Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux

  and same for 4.15.0-29-generic and 4.17.0-8-generic

  Steps to reproduce this bug:

  git clone git://kernel.ubuntu.com/cking/stress-ng
  cd stress-ng
  make clean
  make

  And run with:

  ./stress-ng --sysfs 0 -t 60

  .. wait a few seconds and then:

  [  119.445891] [ cut here ]
  [  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP
  [  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
  [  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO  
   4.15.0-33-generic #36-Ubuntu
  [  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
  [  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
  [  119.446198] 00690648 7cc3de40 
7a74b000
  [  119.446202]03ff80008210  03ff800058da 
7ac1bce8
  [  119.446210] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
  [  119.446210]03ff80005918: c0f40560  brcl
15,3ff800063d8
  [  119.446210]   #03ff8000591e: a7f40001  brc 
15,3ff80005920
  [  119.446210]   >03ff80005922: 0707  bcr 0,%r7
  [  119.446210]03ff80005924: 0707  bcr 0,%r7
  [  119.446210]03ff80005926: 0707  bcr 0,%r7
  [  119.446210]03ff80005928: c004  brcl
0,3ff80005928
  [  119.446210]03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [  119.446226] Call Trace:
  [  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
  [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80
  [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8
  [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8
  [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150
  [  119.446274]  [<00382786>] SyS_read+0x66/0xe0
  [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8
  [  119.446279] Last Breaking-Event-Address:
  [  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
  [  119.446283]
  [  119.446284] ---[ end trace 2c2403d726047e4a ]---

  For  4.17.0-8-generic:
  [   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
  [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP
  [   25.170797] Modules linked in: lttng_statedump(OE) lttng_clock(OE) 
lttng_lib_ring_buffer(OE) binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) 
isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm nfsd 
ib_core auth_rpcgss iscsi_tcp nfs_acl lockd grace libiscsi_tcp libiscsi 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear virtio_net virtio_blk crc32_vx_s390
  [   25.170835] CPU: 0 PID: 5590 Comm: stress-ng-sysfs Tainted: P   OE 
4.17.0-8-generic #9-Ubuntu
  [   25.170837] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [   25.170839] Krnl PSW : 05f0c968 26542d57 
(virtblk_cache_type_show+0x7c/0x80 [virtio_blk])
  [   25.170846]R:0 T:1 IO:1 EX:1 

[Kernel-packages] [Bug 1789638] Re: azure 4.15 kernel: reading sysfs file causing oops

2018-08-31 Thread Colin Ian King
The above patch fixes the issue for me. Thanks

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1789638

Title:
  azure 4.15 kernel: reading sysfs file causing oops

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Kernel: 4.15.0-1021-azure, in Xenial VM on Azure.

  How to reproduce:

  git clone git://kernel.ubuntu.com/cking/stress-ng
  cd stress-ng
  make
  ./stress-ng --sysfs 0 -t 120

  One gets the following:

  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
  [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
  [   22.455286] CPU: 1 PID: 1670 Comm: cat Not tainted 4.15.0-1021-azure 
#21~16.04.1-Ubuntu
  [   22.455286] Hardware name: Microsoft Corporation Virtual Machine/Virtual 
Machine, BIOS 090007  06/02/2017
  [   22.455286] RIP: 0010:read_avail_show+0x1c/0x40
  [   22.455286] RSP: 0018:afa4c4eafdb0 EFLAGS: 00010286
  [   22.455286] RAX:  RBX: 9db36c93e880 RCX: 
9db36f136908
  [   22.860062] RDX:  RSI: 9db364548000 RDI: 
9db364548000
  [   22.888042] RBP: afa4c4eafdb0 R08: 9db364548000 R09: 
9db36c049840
  [   22.920041] R10: 9db364548000 R11:  R12: 
92ae9440
  [   22.948058] R13: 9db36c22d200 R14: 0001 R15: 
9db36c93e880
  [   22.972043] FS:  7f67eeec6700() GS:9db37fd0() 
knlGS:
  [   23.004046] CS:  0010 DS:  ES:  CR0: 80050033
  [   23.024016] CR2: 0004 CR3: 00042c37a003 CR4: 
001606e0
  [   23.048014] Call Trace:
  [   23.060019]  vmbus_chan_attr_show+0x21/0x30
  [   23.076018]  sysfs_kf_seq_show+0xa2/0x130
  [   23.088030]  kernfs_seq_show+0x27/0x30
  [   23.100020]  seq_read+0xb7/0x480
  [   23.112014]  kernfs_fop_read+0x111/0x190
  [   23.128017]  ? security_file_permission+0xa1/0xc0
  [   23.144013]  __vfs_read+0x1b/0x40
  [   23.156019]  vfs_read+0x93/0x130
  [   23.168013]  SyS_read+0x55/0xc0
  [   23.180021]  do_syscall_64+0x73/0x130
  [   23.192014]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [   23.212022] RIP: 0033:0x7f67ee9d8260
  [   23.224016] RSP: 002b:7fffdc193ff8 EFLAGS: 0246 ORIG_RAX: 

  [   23.252022] RAX: ffda RBX: 0002 RCX: 
7f67ee9d8260
  [   23.276019] RDX: 0002 RSI: 7f67eed0c000 RDI: 
0003
  [   23.300020] RBP: 0002 R08:  R09: 

  [   23.328025] R10: 037b R11: 0246 R12: 
7f67eed0c000
  [   23.352036] R13: 0003 R14:  R15: 
0002
  [   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6
  [   23.444022] RIP: read_avail_show+0x1c/0x40 RSP: afa4c4eafdb0
  [   23.468021] CR2: 0004
  [   23.481135] ---[ end trace 348a4b7d5a6747d1 ]---

  Cornered this down to just reading:

  cat
  /sys/devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/99221fa0
  -24ad-11e2-be98-001aa01bbf6e/channels/4/read_avail

  There are various /sysfs VMBUS files that trigger this, see a fix on
  comment #5 below that addresses all the ones I could find.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1789638/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1788432] Re: 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-30 Thread Colin Ian King
Host information:

lsb_release  -a
No LSB modules are available.
Distributor ID: Ubuntu
Description:Ubuntu 16.04.4 LTS
Release:16.04
Codename:   xenial

uname -a
Linux s2lp5 4.4.0-130-generic #156-Ubuntu SMP Thu Jun 14 09:18:59 UTC 2018 
s390x s390x s390x GNU/Linux

qemu: 2.5+dfsg-5ubuntu10.30

this was created using uvt:

#!/bin/bash
#
# Quick and dirty script to create temporary guests which are reachable
# from the kernel VPN via uvt-kvm. After creation one has to login once
# via "virsh console " to figure out the ip address. :/

BASEDIR=$(dirname $0)

if [ "$1" = "" ]; then
echo "$(basename $0) "
exit 1
fi
VMNAME="$1"
shift
uvt-kvm create --cpu 2 --memory 2048 --disk 20 --password ubuntu \
--template $BASEDIR/uvt-template.xml $VMNAME arch=s390x "$@"


uvt-template.xml:



hvm

1
destroy
restart
preserve

/usr/bin/qemu-system-s390x










Hope that's enough info

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1788432

Title:
  4.15 s390x kernel BUG at /build/linux-
  Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

Status in Ubuntu on IBM z Systems:
  Triaged
Status in linux package in Ubuntu:
  Triaged
Status in linux source package in Bionic:
  Triaged

Bug description:
  uname -a
  Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux

  and same for 4.15.0-29-generic and 4.17.0-8-generic

  Steps to reproduce this bug:

  git clone git://kernel.ubuntu.com/cking/stress-ng
  cd stress-ng
  make clean
  make

  And run with:

  ./stress-ng --sysfs 0 -t 60

  .. wait a few seconds and then:

  [  119.445891] [ cut here ]
  [  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP
  [  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
  [  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO  
   4.15.0-33-generic #36-Ubuntu
  [  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
  [  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
  [  119.446198] 00690648 7cc3de40 
7a74b000
  [  119.446202]03ff80008210  03ff800058da 
7ac1bce8
  [  119.446210] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
  [  119.446210]03ff80005918: c0f40560  brcl
15,3ff800063d8
  [  119.446210]   #03ff8000591e: a7f40001  brc 
15,3ff80005920
  [  119.446210]   >03ff80005922: 0707  bcr 0,%r7
  [  119.446210]03ff80005924: 0707  bcr 0,%r7
  [  119.446210]03ff80005926: 0707  bcr 0,%r7
  [  119.446210]03ff80005928: c004  brcl
0,3ff80005928
  [  119.446210]03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [  119.446226] Call Trace:
  [  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
  [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80
  [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8
  [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8
  [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150
  [  119.446274]  [<00382786>] SyS_read+0x66/0xe0
  [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8
  [  119.446279] Last Breaking-Event-Address:
  [  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
  [  119.446283]
  [  119.446284] ---[ end trace 2c2403d726047e4a ]---

  For  4.17.0-8-generic:
  [   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
  [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP
  [   25.170797] Modules linked in: 

[Kernel-packages] [Bug 1789638] Re: azure 4.15 kernel: reading sysfs file causing oops

2018-08-29 Thread Colin Ian King
Workaround fix attached. I suspect there may be a more elegant solution.


** Patch added: "workaround fix"
   
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1789638/+attachment/5182213/+files/0001-vmbus-check-for-null-ring-buffer-before-attempting-t.patch

** Description changed:

  Kernel: 4.15.0-1021-azure
  
  How to reproduce:
  
- stress-ng --sysfs 0 -t 120
+ git clone git://kernel.ubuntu.com/cking/stress-ng
+ cd stress-ng
+ make
+ ./stress-ng --sysfs 0 -t 120
  
  One gets the following:
  
  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
- [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0 
+ [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
  [   22.455286] CPU: 1 PID: 1670 Comm: cat Not tainted 4.15.0-1021-azure 
#21~16.04.1-Ubuntu
  [   22.455286] Hardware name: Microsoft Corporation Virtual Machine/Virtual 
Machine, BIOS 090007  06/02/2017
  [   22.455286] RIP: 0010:read_avail_show+0x1c/0x40
  [   22.455286] RSP: 0018:afa4c4eafdb0 EFLAGS: 00010286
  [   22.455286] RAX:  RBX: 9db36c93e880 RCX: 
9db36f136908
  [   22.860062] RDX:  RSI: 9db364548000 RDI: 
9db364548000
  [   22.888042] RBP: afa4c4eafdb0 R08: 9db364548000 R09: 
9db36c049840
  [   22.920041] R10: 9db364548000 R11:  R12: 
92ae9440
  [   22.948058] R13: 9db36c22d200 R14: 0001 R15: 
9db36c93e880
  [   22.972043] FS:  7f67eeec6700() GS:9db37fd0() 
knlGS:
  [   23.004046] CS:  0010 DS:  ES:  CR0: 80050033
  [   23.024016] CR2: 0004 CR3: 00042c37a003 CR4: 
001606e0
  [   23.048014] Call Trace:
  [   23.060019]  vmbus_chan_attr_show+0x21/0x30
  [   23.076018]  sysfs_kf_seq_show+0xa2/0x130
  [   23.088030]  kernfs_seq_show+0x27/0x30
  [   23.100020]  seq_read+0xb7/0x480
  [   23.112014]  kernfs_fop_read+0x111/0x190
  [   23.128017]  ? security_file_permission+0xa1/0xc0
  [   23.144013]  __vfs_read+0x1b/0x40
  [   23.156019]  vfs_read+0x93/0x130
  [   23.168013]  SyS_read+0x55/0xc0
  [   23.180021]  do_syscall_64+0x73/0x130
  [   23.192014]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [   23.212022] RIP: 0033:0x7f67ee9d8260
  [   23.224016] RSP: 002b:7fffdc193ff8 EFLAGS: 0246 ORIG_RAX: 

  [   23.252022] RAX: ffda RBX: 0002 RCX: 
7f67ee9d8260
  [   23.276019] RDX: 0002 RSI: 7f67eed0c000 RDI: 
0003
  [   23.300020] RBP: 0002 R08:  R09: 

  [   23.328025] R10: 037b R11: 0246 R12: 
7f67eed0c000
  [   23.352036] R13: 0003 R14:  R15: 
0002
- [   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6 
+ [   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6
  [   23.444022] RIP: read_avail_show+0x1c/0x40 RSP: afa4c4eafdb0
  [   23.468021] CR2: 0004
  [   23.481135] ---[ end trace 348a4b7d5a6747d1 ]---
  
  Cornered this down to just reading:
  
  cat
  /sys/devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/99221fa0
  -24ad-11e2-be98-001aa01bbf6e/channels/4/read_avail
+ 
+ There are various /sysfs VMBUS files that trigger this, see a fix on
+ comment #5 below that addresses all the ones I could find.

** Description changed:

- Kernel: 4.15.0-1021-azure
+ Kernel: 4.15.0-1021-azure, in Xenial VM on Azure.
  
  How to reproduce:
  
  git clone git://kernel.ubuntu.com/cking/stress-ng
  cd stress-ng
  make
  ./stress-ng --sysfs 0 -t 120
  
  One gets the following:
  
  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
  [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules 

[Kernel-packages] [Bug 1789638] Re: azure 4.15 kernel: reading sysfs file causing oops

2018-08-29 Thread Colin Ian King
Same issue with rbi->ring_buffer on write_avail_show() too.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1789638

Title:
  azure 4.15 kernel: reading sysfs file causing oops

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Kernel: 4.15.0-1021-azure

  How to reproduce:

  stress-ng --sysfs 0 -t 120

  One gets the following:

  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
  [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0 
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
  [   22.455286] CPU: 1 PID: 1670 Comm: cat Not tainted 4.15.0-1021-azure 
#21~16.04.1-Ubuntu
  [   22.455286] Hardware name: Microsoft Corporation Virtual Machine/Virtual 
Machine, BIOS 090007  06/02/2017
  [   22.455286] RIP: 0010:read_avail_show+0x1c/0x40
  [   22.455286] RSP: 0018:afa4c4eafdb0 EFLAGS: 00010286
  [   22.455286] RAX:  RBX: 9db36c93e880 RCX: 
9db36f136908
  [   22.860062] RDX:  RSI: 9db364548000 RDI: 
9db364548000
  [   22.888042] RBP: afa4c4eafdb0 R08: 9db364548000 R09: 
9db36c049840
  [   22.920041] R10: 9db364548000 R11:  R12: 
92ae9440
  [   22.948058] R13: 9db36c22d200 R14: 0001 R15: 
9db36c93e880
  [   22.972043] FS:  7f67eeec6700() GS:9db37fd0() 
knlGS:
  [   23.004046] CS:  0010 DS:  ES:  CR0: 80050033
  [   23.024016] CR2: 0004 CR3: 00042c37a003 CR4: 
001606e0
  [   23.048014] Call Trace:
  [   23.060019]  vmbus_chan_attr_show+0x21/0x30
  [   23.076018]  sysfs_kf_seq_show+0xa2/0x130
  [   23.088030]  kernfs_seq_show+0x27/0x30
  [   23.100020]  seq_read+0xb7/0x480
  [   23.112014]  kernfs_fop_read+0x111/0x190
  [   23.128017]  ? security_file_permission+0xa1/0xc0
  [   23.144013]  __vfs_read+0x1b/0x40
  [   23.156019]  vfs_read+0x93/0x130
  [   23.168013]  SyS_read+0x55/0xc0
  [   23.180021]  do_syscall_64+0x73/0x130
  [   23.192014]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [   23.212022] RIP: 0033:0x7f67ee9d8260
  [   23.224016] RSP: 002b:7fffdc193ff8 EFLAGS: 0246 ORIG_RAX: 

  [   23.252022] RAX: ffda RBX: 0002 RCX: 
7f67ee9d8260
  [   23.276019] RDX: 0002 RSI: 7f67eed0c000 RDI: 
0003
  [   23.300020] RBP: 0002 R08:  R09: 

  [   23.328025] R10: 037b R11: 0246 R12: 
7f67eed0c000
  [   23.352036] R13: 0003 R14:  R15: 
0002
  [   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6 
  [   23.444022] RIP: read_avail_show+0x1c/0x40 RSP: afa4c4eafdb0
  [   23.468021] CR2: 0004
  [   23.481135] ---[ end trace 348a4b7d5a6747d1 ]---

  Cornered this down to just reading:

  cat
  /sys/devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/99221fa0
  -24ad-11e2-be98-001aa01bbf6e/channels/4/read_avail

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1789638/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1789638] Re: azure 4.15 kernel: reading sysfs file causing oops

2018-08-29 Thread Colin Ian King
Tested with today's linux tip, 4.19.0-rc1 @ commit
3f16503b7d2274ac8cbab11163047ac0b4c66cfe, issue still occurs.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1789638

Title:
  azure 4.15 kernel: reading sysfs file causing oops

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Kernel: 4.15.0-1021-azure

  How to reproduce:

  stress-ng --sysfs 0 -t 120

  One gets the following:

  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
  [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0 
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
  [   22.455286] CPU: 1 PID: 1670 Comm: cat Not tainted 4.15.0-1021-azure 
#21~16.04.1-Ubuntu
  [   22.455286] Hardware name: Microsoft Corporation Virtual Machine/Virtual 
Machine, BIOS 090007  06/02/2017
  [   22.455286] RIP: 0010:read_avail_show+0x1c/0x40
  [   22.455286] RSP: 0018:afa4c4eafdb0 EFLAGS: 00010286
  [   22.455286] RAX:  RBX: 9db36c93e880 RCX: 
9db36f136908
  [   22.860062] RDX:  RSI: 9db364548000 RDI: 
9db364548000
  [   22.888042] RBP: afa4c4eafdb0 R08: 9db364548000 R09: 
9db36c049840
  [   22.920041] R10: 9db364548000 R11:  R12: 
92ae9440
  [   22.948058] R13: 9db36c22d200 R14: 0001 R15: 
9db36c93e880
  [   22.972043] FS:  7f67eeec6700() GS:9db37fd0() 
knlGS:
  [   23.004046] CS:  0010 DS:  ES:  CR0: 80050033
  [   23.024016] CR2: 0004 CR3: 00042c37a003 CR4: 
001606e0
  [   23.048014] Call Trace:
  [   23.060019]  vmbus_chan_attr_show+0x21/0x30
  [   23.076018]  sysfs_kf_seq_show+0xa2/0x130
  [   23.088030]  kernfs_seq_show+0x27/0x30
  [   23.100020]  seq_read+0xb7/0x480
  [   23.112014]  kernfs_fop_read+0x111/0x190
  [   23.128017]  ? security_file_permission+0xa1/0xc0
  [   23.144013]  __vfs_read+0x1b/0x40
  [   23.156019]  vfs_read+0x93/0x130
  [   23.168013]  SyS_read+0x55/0xc0
  [   23.180021]  do_syscall_64+0x73/0x130
  [   23.192014]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [   23.212022] RIP: 0033:0x7f67ee9d8260
  [   23.224016] RSP: 002b:7fffdc193ff8 EFLAGS: 0246 ORIG_RAX: 

  [   23.252022] RAX: ffda RBX: 0002 RCX: 
7f67ee9d8260
  [   23.276019] RDX: 0002 RSI: 7f67eed0c000 RDI: 
0003
  [   23.300020] RBP: 0002 R08:  R09: 

  [   23.328025] R10: 037b R11: 0246 R12: 
7f67eed0c000
  [   23.352036] R13: 0003 R14:  R15: 
0002
  [   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6 
  [   23.444022] RIP: read_avail_show+0x1c/0x40 RSP: afa4c4eafdb0
  [   23.468021] CR2: 0004
  [   23.481135] ---[ end trace 348a4b7d5a6747d1 ]---

  Cornered this down to just reading:

  cat
  /sys/devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/99221fa0
  -24ad-11e2-be98-001aa01bbf6e/channels/4/read_avail

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1789638/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1789638] Re: azure 4.15 kernel: reading sysfs file causing oops

2018-08-29 Thread Colin Ian King
In read_avail_show, rbi->ring_buffer is NULL, causing the OOPS.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1789638

Title:
  azure 4.15 kernel: reading sysfs file causing oops

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Kernel: 4.15.0-1021-azure

  How to reproduce:

  stress-ng --sysfs 0 -t 120

  One gets the following:

  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
  [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0 
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
  [   22.455286] CPU: 1 PID: 1670 Comm: cat Not tainted 4.15.0-1021-azure 
#21~16.04.1-Ubuntu
  [   22.455286] Hardware name: Microsoft Corporation Virtual Machine/Virtual 
Machine, BIOS 090007  06/02/2017
  [   22.455286] RIP: 0010:read_avail_show+0x1c/0x40
  [   22.455286] RSP: 0018:afa4c4eafdb0 EFLAGS: 00010286
  [   22.455286] RAX:  RBX: 9db36c93e880 RCX: 
9db36f136908
  [   22.860062] RDX:  RSI: 9db364548000 RDI: 
9db364548000
  [   22.888042] RBP: afa4c4eafdb0 R08: 9db364548000 R09: 
9db36c049840
  [   22.920041] R10: 9db364548000 R11:  R12: 
92ae9440
  [   22.948058] R13: 9db36c22d200 R14: 0001 R15: 
9db36c93e880
  [   22.972043] FS:  7f67eeec6700() GS:9db37fd0() 
knlGS:
  [   23.004046] CS:  0010 DS:  ES:  CR0: 80050033
  [   23.024016] CR2: 0004 CR3: 00042c37a003 CR4: 
001606e0
  [   23.048014] Call Trace:
  [   23.060019]  vmbus_chan_attr_show+0x21/0x30
  [   23.076018]  sysfs_kf_seq_show+0xa2/0x130
  [   23.088030]  kernfs_seq_show+0x27/0x30
  [   23.100020]  seq_read+0xb7/0x480
  [   23.112014]  kernfs_fop_read+0x111/0x190
  [   23.128017]  ? security_file_permission+0xa1/0xc0
  [   23.144013]  __vfs_read+0x1b/0x40
  [   23.156019]  vfs_read+0x93/0x130
  [   23.168013]  SyS_read+0x55/0xc0
  [   23.180021]  do_syscall_64+0x73/0x130
  [   23.192014]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
  [   23.212022] RIP: 0033:0x7f67ee9d8260
  [   23.224016] RSP: 002b:7fffdc193ff8 EFLAGS: 0246 ORIG_RAX: 

  [   23.252022] RAX: ffda RBX: 0002 RCX: 
7f67ee9d8260
  [   23.276019] RDX: 0002 RSI: 7f67eed0c000 RDI: 
0003
  [   23.300020] RBP: 0002 R08:  R09: 

  [   23.328025] R10: 037b R11: 0246 R12: 
7f67eed0c000
  [   23.352036] R13: 0003 R14:  R15: 
0002
  [   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6 
  [   23.444022] RIP: read_avail_show+0x1c/0x40 RSP: afa4c4eafdb0
  [   23.468021] CR2: 0004
  [   23.481135] ---[ end trace 348a4b7d5a6747d1 ]---

  Cornered this down to just reading:

  cat
  /sys/devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/99221fa0
  -24ad-11e2-be98-001aa01bbf6e/channels/4/read_avail

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1789638/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1789638] [NEW] azure 4.15 kernel: reading sysfs file causing oops

2018-08-29 Thread Colin Ian King
Public bug reported:

Kernel: 4.15.0-1021-azure

How to reproduce:

stress-ng --sysfs 0 -t 120

One gets the following:

[   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
[   22.455286] IP: read_avail_show+0x1c/0x40
[   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0 
[   22.455286] Oops:  [#1] SMP PTI
[   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
[   22.455286] CPU: 1 PID: 1670 Comm: cat Not tainted 4.15.0-1021-azure 
#21~16.04.1-Ubuntu
[   22.455286] Hardware name: Microsoft Corporation Virtual Machine/Virtual 
Machine, BIOS 090007  06/02/2017
[   22.455286] RIP: 0010:read_avail_show+0x1c/0x40
[   22.455286] RSP: 0018:afa4c4eafdb0 EFLAGS: 00010286
[   22.455286] RAX:  RBX: 9db36c93e880 RCX: 9db36f136908
[   22.860062] RDX:  RSI: 9db364548000 RDI: 9db364548000
[   22.888042] RBP: afa4c4eafdb0 R08: 9db364548000 R09: 9db36c049840
[   22.920041] R10: 9db364548000 R11:  R12: 92ae9440
[   22.948058] R13: 9db36c22d200 R14: 0001 R15: 9db36c93e880
[   22.972043] FS:  7f67eeec6700() GS:9db37fd0() 
knlGS:
[   23.004046] CS:  0010 DS:  ES:  CR0: 80050033
[   23.024016] CR2: 0004 CR3: 00042c37a003 CR4: 001606e0
[   23.048014] Call Trace:
[   23.060019]  vmbus_chan_attr_show+0x21/0x30
[   23.076018]  sysfs_kf_seq_show+0xa2/0x130
[   23.088030]  kernfs_seq_show+0x27/0x30
[   23.100020]  seq_read+0xb7/0x480
[   23.112014]  kernfs_fop_read+0x111/0x190
[   23.128017]  ? security_file_permission+0xa1/0xc0
[   23.144013]  __vfs_read+0x1b/0x40
[   23.156019]  vfs_read+0x93/0x130
[   23.168013]  SyS_read+0x55/0xc0
[   23.180021]  do_syscall_64+0x73/0x130
[   23.192014]  entry_SYSCALL_64_after_hwframe+0x3d/0xa2
[   23.212022] RIP: 0033:0x7f67ee9d8260
[   23.224016] RSP: 002b:7fffdc193ff8 EFLAGS: 0246 ORIG_RAX: 

[   23.252022] RAX: ffda RBX: 0002 RCX: 7f67ee9d8260
[   23.276019] RDX: 0002 RSI: 7f67eed0c000 RDI: 0003
[   23.300020] RBP: 0002 R08:  R09: 
[   23.328025] R10: 037b R11: 0246 R12: 7f67eed0c000
[   23.352036] R13: 0003 R14:  R15: 0002
[   23.376678] Code: fb 3a 17 00 48 98 5d c3 0f 1f 80 00 00 00 00 0f 1f 44 00 
00 55 48 8b 87 38 01 00 00 49 89 f0 8b 97 48 01 00 00 4c 89 c7 48 89 e5 <8b> 48 
04 8b 00 29 ca 89 c6 29 ce 01 c2 39 c1 0f 46 d6 48 c7 c6 
[   23.444022] RIP: read_avail_show+0x1c/0x40 RSP: afa4c4eafdb0
[   23.468021] CR2: 0004
[   23.481135] ---[ end trace 348a4b7d5a6747d1 ]---

Cornered this down to just reading:

cat
/sys/devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/99221fa0
-24ad-11e2-be98-001aa01bbf6e/channels/4/read_avail

** Affects: linux (Ubuntu)
 Importance: Medium
 Status: Incomplete


** Tags: bionic

** Changed in: linux (Ubuntu)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1789638

Title:
  azure 4.15 kernel: reading sysfs file causing oops

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Kernel: 4.15.0-1021-azure

  How to reproduce:

  stress-ng --sysfs 0 -t 120

  One gets the following:

  [   22.451885] BUG: unable to handle kernel NULL pointer dereference at 
0004
  [   22.455286] IP: read_avail_show+0x1c/0x40
  [   22.455286] PGD 80042d59e067 P4D 80042d59e067 PUD 42eb8c067 PMD 0 
  [   22.455286] Oops:  [#1] SMP PTI
  [   22.455286] Modules linked in: nf_conntrack_ipv4 nf_defrag_ipv4 xt_owner 
xt_conntrack nf_conntrack iptable_security ip_tables x_tables serio_raw joydev 
hv_balloon ib_iser iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi autofs4 
btrfs zstd_compress raid10 raid456 async_raid6_recov async_memcpy async_pq 
async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear 
hid_generic crct10dif_pclmul crc32_pclmul ghash_clmulni_intel pcbc aesni_intel 
aes_x86_64 crypto_simd glue_helper cryptd hyperv_fb hid_hyperv pata_acpi 
cfbfillrect hyperv_keyboard cfbimgblt hid cfbcopyarea hv_netvsc hv_utils
  [   22.455286] CPU: 1 PID: 1670 Comm: 

[Kernel-packages] [Bug 1757057] Re: sysfs test in ubuntu_stress_smoke will induce kernel oops on X-HWE ppc64le

2018-08-29 Thread Colin Ian King
Yes, I believe that is true, the stressors now try to exercise a wider
set of sysfs files in a more efficient way, so I expect we're hitting
pre existing issues.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1757057

Title:
  sysfs test in ubuntu_stress_smoke will induce kernel oops on X-HWE
  ppc64le

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Proposed kernel 4.13.0-38.43~16.04.1 on ppc64le

  sysfs FAILED (kernel oopsed)
  [ 891.312202] WARNING! power/level is deprecated; use power/control instead
  [ 892.138260] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.138627] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.138976] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.139303] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.139610] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 255
  [ 892.141571] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.141697] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142003] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142319] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142616] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.142903] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 87
  [ 892.148239] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.148583] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.149406] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.149530] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 255
  [ 892.149855] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.150161] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.150467] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.151980] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.152255] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.152558] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.152980] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.153545] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.153908] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.154197] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.154502] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.154813] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.155110] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.155407] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.162052] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.162406] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.162708] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 85
  [ 892.163007] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.163304] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.163612] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.163912] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.164505] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.164816] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.165118] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.165413] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.165717] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166004] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166309] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166613] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.166908] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.167212] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.167524] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.167803] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 22
  [ 892.168128] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.168439] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.168732] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169038] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169335] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169630] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.169934] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.170235] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176000] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176350] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176643] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.176939] ses 0:0:8:0: Wrong diagnostic page; asked for 2 got 46
  [ 892.177258] ses 0:0:8:0: 

[Kernel-packages] [Bug 1788432] Re: 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-28 Thread Colin Ian King
** Description changed:

  uname -a
  Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux
  
  and same for 4.15.0-29-generic and 4.17.0-8-generic
+ 
+ Steps to reproduce this bug:
+ 
+ git clone git://kernel.ubuntu.com/cking/stress-ng
+ cd stress-ng
+ make clean
+ make
+ 
+ And run with:
  
  ./stress-ng --sysfs 0 -t 60
  
  .. wait a few seconds and then:
  
  [  119.445891] [ cut here ]
  [  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
- [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP 
+ [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP
  [  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
  [  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO  
   4.15.0-33-generic #36-Ubuntu
  [  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
  [  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
  [  119.446198] 00690648 7cc3de40 
7a74b000
  [  119.446202]03ff80008210  03ff800058da 
7ac1bce8
  [  119.446210] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
  [  119.446210]03ff80005918: c0f40560  brcl
15,3ff800063d8
  [  119.446210]   #03ff8000591e: a7f40001  brc 
15,3ff80005920
  [  119.446210]   >03ff80005922: 0707  bcr 0,%r7
  [  119.446210]03ff80005924: 0707  bcr 0,%r7
  [  119.446210]03ff80005926: 0707  bcr 0,%r7
  [  119.446210]03ff80005928: c004  brcl
0,3ff80005928
  [  119.446210]03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [  119.446226] Call Trace:
  [  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
- [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80 
- [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8 
- [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8 
- [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150 
- [  119.446274]  [<00382786>] SyS_read+0x66/0xe0 
- [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8 
+ [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80
+ [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8
+ [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8
+ [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150
+ [  119.446274]  [<00382786>] SyS_read+0x66/0xe0
+ [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8
  [  119.446279] Last Breaking-Event-Address:
  [  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
- [  119.446283]  
+ [  119.446283]
  [  119.446284] ---[ end trace 2c2403d726047e4a ]---
- 
  
  For  4.17.0-8-generic:
  [   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
- [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP 
+ [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP
  [   25.170797] Modules linked in: lttng_statedump(OE) lttng_clock(OE) 
lttng_lib_ring_buffer(OE) binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) 
isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm nfsd 
ib_core auth_rpcgss iscsi_tcp nfs_acl lockd grace libiscsi_tcp libiscsi 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear virtio_net virtio_blk crc32_vx_s390
  [   25.170835] CPU: 0 PID: 5590 Comm: stress-ng-sysfs Tainted: P   OE 
4.17.0-8-generic #9-Ubuntu
  [   25.170837] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [   25.170839] Krnl PSW : 05f0c968 26542d57 
(virtblk_cache_type_show+0x7c/0x80 [virtio_blk])

[Kernel-packages] [Bug 1788432] Re: 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-28 Thread Colin Ian King
I think this is a race in ccw_io_helper(). The stress-ng sysfs stressor
is running multiple threaded reads of
/sys/devices/css0/0.0./0.0./virtio0/block/vda/cache_type which
results in reads via virtio_cread8() and ultimately reads using
virtio_ccw_get_config()

I added debug into virtio_cread8 and it returns 1 99% of the time, and
sometimes in a threaded read we get it returning garbage, such as 200 in
the following trace:

[   39.76] virtio_cread8 32 -> 1
[   39.767933] virtio_cread8 32 -> 1
[   39.787712] virtio_cread8 32 -> 200
[   39.787810] kernel BUG at drivers/block/virtio_blk.c:576!

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1788432

Title:
  4.15 s390x kernel BUG at /build/linux-
  Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

Status in linux package in Ubuntu:
  Triaged
Status in linux source package in Bionic:
  Triaged

Bug description:
  uname -a
  Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux

  and same for 4.15.0-29-generic and 4.17.0-8-generic

  ./stress-ng --sysfs 0 -t 60

  .. wait a few seconds and then:

  [  119.445891] [ cut here ]
  [  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP 
  [  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
  [  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO  
   4.15.0-33-generic #36-Ubuntu
  [  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
  [  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
  [  119.446198] 00690648 7cc3de40 
7a74b000
  [  119.446202]03ff80008210  03ff800058da 
7ac1bce8
  [  119.446210] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
  [  119.446210]03ff80005918: c0f40560  brcl
15,3ff800063d8
  [  119.446210]   #03ff8000591e: a7f40001  brc 
15,3ff80005920
  [  119.446210]   >03ff80005922: 0707  bcr 0,%r7
  [  119.446210]03ff80005924: 0707  bcr 0,%r7
  [  119.446210]03ff80005926: 0707  bcr 0,%r7
  [  119.446210]03ff80005928: c004  brcl
0,3ff80005928
  [  119.446210]03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [  119.446226] Call Trace:
  [  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
  [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80 
  [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8 
  [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8 
  [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150 
  [  119.446274]  [<00382786>] SyS_read+0x66/0xe0 
  [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8 
  [  119.446279] Last Breaking-Event-Address:
  [  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
  [  119.446283]  
  [  119.446284] ---[ end trace 2c2403d726047e4a ]---

  
  For  4.17.0-8-generic:
  [   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
  [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP 
  [   25.170797] Modules linked in: lttng_statedump(OE) lttng_clock(OE) 
lttng_lib_ring_buffer(OE) binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) 
isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm nfsd 
ib_core auth_rpcgss iscsi_tcp nfs_acl lockd grace libiscsi_tcp libiscsi 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear 

[Kernel-packages] [Bug 1783651] Re: Please enable CONFIG_PAGE_POISONING

2018-08-24 Thread Colin Ian King
Patch sent to kernel team mailing list

** Changed in: linux (Ubuntu Cosmic)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: linux (Ubuntu Cosmic)
   Status: Confirmed => In Progress

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1783651

Title:
  Please enable CONFIG_PAGE_POISONING

Status in linux package in Ubuntu:
  In Progress
Status in linux source package in Cosmic:
  In Progress

Bug description:
  I'd like to be able to use page poisoning, but CONFIG_PAGE_POISONING
  is not enabled on Ubuntu. (This option itself has a near-zero
  performance impact since it must be combined with the boot option
  "page_poison=1" to actually enable the poisoning.)

  To make the poisoning (when enabled) less of an impact, I'd also like
  to see CONFIG_PAGE_POISONING_ZERO=y too, which means GFP_ZEROing can
  be skipped.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1783651/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1788385] Re: vfork in ubuntu_stress_smoke test failed on KVM node with B-KVM kernel

2018-08-23 Thread Colin Ian King
This is not a kernel error; it's a bug in stress-ng when running with
the --verify option on systems where memory is low and EAGAIN or ENOMEM
occurs.  I've fixed this in stress-ng with commit:

http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=8ab0ac7ce995517c0cb10db60bee5a19b7b8

Do you mind running a re-test to see if this fixes the issue.  Apologies
for this stress-ng bug. My bad.

** Changed in: stress-ng
   Importance: Undecided => High

** Changed in: stress-ng
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: stress-ng
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1788385

Title:
  vfork in ubuntu_stress_smoke test failed on KVM node with B-KVM kernel

Status in Stress-ng:
  In Progress
Status in ubuntu-kernel-tests:
  New
Status in linux-kvm package in Ubuntu:
  New

Bug description:
  Fail rate 2 out of 2, failed on a kvm node "gummo", with 1G ram

  uname -vr : 4.15.0-1020-kvm #20-Ubuntu SMP Thu Aug 16 16:46:26 UTC 2018
  Bionic 18.04
  arch: amd64

  DEBUG - 11:24:24 DEBUG| [stdout] vfork STARTING
  DEBUG - 11:24:24 DEBUG| [stdout] vfork STARTING
  DEBUG - 11:24:26 DEBUG| [stdout] vfork RETURNED 2
  DEBUG - 11:24:26 DEBUG| [stdout] vfork RETURNED 2
  DEBUG - 11:24:26 DEBUG| [stdout] vfork FAILED
  DEBUG - 11:24:26 DEBUG| [stdout] vfork FAILED
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 2 
processors online, 2 processors configured
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 2 
processors online, 2 processors configured
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: info:  [1437] 
dispatching hogs: 4 vfork
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: info:  [1437] 
dispatching hogs: 4 vfork
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: reducing cache level from L3 (too high) to L2
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: reducing cache level from L3 (too high) to L2
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: default cache size: 4096K
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: default cache size: 4096K
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 
starting stressors
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 
starting stressors
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1438] 
stress-ng-vfork: started [1438] (instance 0)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1438] 
stress-ng-vfork: started [1438] (instance 0)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 4 
stressors spawned
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 4 
stressors spawned
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1439] 
stress-ng-vfork: started [1439] (instance 1)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1439] 
stress-ng-vfork: started [1439] (instance 1)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1440] 
stress-ng-vfork: started [1440] (instance 2)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1440] 
stress-ng-vfork: started [1440] (instance 2)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1441] 
stress-ng-vfork: started [1441] (instance 3)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1441] 
stress-ng-vfork: started [1441] (instance 3)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1438] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1438] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [144

[Kernel-packages] [Bug 1788432] Re: 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-23 Thread Colin Ian King
FYI, I've added easier detection of which sysfs file is causing this
with commit: http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=690b6392645219702f410318a3c99bee7c4f192b

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1788432

Title:
  4.15 s390x kernel BUG at /build/linux-
  Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

Status in linux package in Ubuntu:
  Triaged
Status in linux source package in Bionic:
  Triaged

Bug description:
  uname -a
  Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux

  and same for 4.15.0-29-generic and 4.17.0-8-generic

  ./stress-ng --sysfs 0 -t 60

  .. wait a few seconds and then:

  [  119.445891] [ cut here ]
  [  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP 
  [  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
  [  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO  
   4.15.0-33-generic #36-Ubuntu
  [  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
  [  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
  [  119.446198] 00690648 7cc3de40 
7a74b000
  [  119.446202]03ff80008210  03ff800058da 
7ac1bce8
  [  119.446210] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
  [  119.446210]03ff80005918: c0f40560  brcl
15,3ff800063d8
  [  119.446210]   #03ff8000591e: a7f40001  brc 
15,3ff80005920
  [  119.446210]   >03ff80005922: 0707  bcr 0,%r7
  [  119.446210]03ff80005924: 0707  bcr 0,%r7
  [  119.446210]03ff80005926: 0707  bcr 0,%r7
  [  119.446210]03ff80005928: c004  brcl
0,3ff80005928
  [  119.446210]03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [  119.446226] Call Trace:
  [  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
  [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80 
  [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8 
  [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8 
  [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150 
  [  119.446274]  [<00382786>] SyS_read+0x66/0xe0 
  [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8 
  [  119.446279] Last Breaking-Event-Address:
  [  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
  [  119.446283]  
  [  119.446284] ---[ end trace 2c2403d726047e4a ]---

  
  For  4.17.0-8-generic:
  [   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
  [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP 
  [   25.170797] Modules linked in: lttng_statedump(OE) lttng_clock(OE) 
lttng_lib_ring_buffer(OE) binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) 
isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm nfsd 
ib_core auth_rpcgss iscsi_tcp nfs_acl lockd grace libiscsi_tcp libiscsi 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear virtio_net virtio_blk crc32_vx_s390
  [   25.170835] CPU: 0 PID: 5590 Comm: stress-ng-sysfs Tainted: P   OE 
4.17.0-8-generic #9-Ubuntu
  [   25.170837] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [   25.170839] Krnl PSW : 05f0c968 26542d57 
(virtblk_cache_type_show+0x7c/0x80 [virtio_blk])
  [   25.170846]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [   25.170849] 

[Kernel-packages] [Bug 1788432] Re: 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-22 Thread Colin Ian King
Maybe related to:
https://objectstorage.prodstack4-5.canonical.com/v1/AUTH_77e2ada1e7a84929a74ba3b87153c0ac
/autopkgtest-bionic/bionic/s390x/l/linux/20180822_104342_a8a5e@/log.gz

** Changed in: linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1788432

Title:
  4.15 s390x kernel BUG at /build/linux-
  Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

Status in linux package in Ubuntu:
  Triaged
Status in linux source package in Bionic:
  Triaged

Bug description:
  uname -a
  Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux

  and same for 4.15.0-29-generic and 4.17.0-8-generic

  ./stress-ng --sysfs 0 -t 60

  .. wait a few seconds and then:

  [  119.445891] [ cut here ]
  [  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [  119.446093] illegal operation: 0001 ilc:1 [#3] SMP 
  [  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
  [  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO  
   4.15.0-33-generic #36-Ubuntu
  [  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
  [  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
  [  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
  [  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
  [  119.446198] 00690648 7cc3de40 
7a74b000
  [  119.446202]03ff80008210  03ff800058da 
7ac1bce8
  [  119.446210] Krnl Code: 03ff80005912: ebbff0a80004  lmg 
%r11,%r15,168(%r15)
  [  119.446210]03ff80005918: c0f40560  brcl
15,3ff800063d8
  [  119.446210]   #03ff8000591e: a7f40001  brc 
15,3ff80005920
  [  119.446210]   >03ff80005922: 0707  bcr 0,%r7
  [  119.446210]03ff80005924: 0707  bcr 0,%r7
  [  119.446210]03ff80005926: 0707  bcr 0,%r7
  [  119.446210]03ff80005928: c004  brcl
0,3ff80005928
  [  119.446210]03ff8000592e: eb6ff0480024  stmg
%r6,%r15,72(%r15)
  [  119.446226] Call Trace:
  [  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
  [  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80 
  [  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8 
  [  119.446259]  [<003b048c>] seq_read+0xec/0x4c8 
  [  119.446262]  [<003821ea>] vfs_read+0x8a/0x150 
  [  119.446274]  [<00382786>] SyS_read+0x66/0xe0 
  [  119.446278]  [<008e3028>] system_call+0xdc/0x2c8 
  [  119.446279] Last Breaking-Event-Address:
  [  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
  [  119.446283]  
  [  119.446284] ---[ end trace 2c2403d726047e4a ]---

  
  For  4.17.0-8-generic:
  [   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
  [   25.170795] illegal operation: 0001 ilc:1 [#1] SMP 
  [   25.170797] Modules linked in: lttng_statedump(OE) lttng_clock(OE) 
lttng_lib_ring_buffer(OE) binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) 
isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm nfsd 
ib_core auth_rpcgss iscsi_tcp nfs_acl lockd grace libiscsi_tcp libiscsi 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear virtio_net virtio_blk crc32_vx_s390
  [   25.170835] CPU: 0 PID: 5590 Comm: stress-ng-sysfs Tainted: P   OE 
4.17.0-8-generic #9-Ubuntu
  [   25.170837] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [   25.170839] Krnl PSW : 05f0c968 26542d57 
(virtblk_c

[Kernel-packages] [Bug 1788314] Re: Conflict between zfs-linux and s390-tools

2018-08-22 Thread Colin Ian King
Thanks for the update on the zfs task.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1788314

Title:
  Conflict between zfs-linux and s390-tools

Status in s390-tools package in Ubuntu:
  Confirmed
Status in zfs-linux package in Ubuntu:
  Invalid

Bug description:
  Not sure which of the two needs fixing, but there's a path conflict
  between zfs-linux and s390-tools which effectively prevents installing
  ZFS on s390x in cosmic.

  (Reading database ... 83042 files and directories currently installed.)
  Preparing to unpack .../zfsutils-linux_0.7.9-3ubuntu5_s390x.deb ...
  Unpacking zfsutils-linux (0.7.9-3ubuntu5) ...
  dpkg: error processing archive 
/var/cache/apt/archives/zfsutils-linux_0.7.9-3ubuntu5_s390x.deb (--unpack):
   trying to overwrite '/usr/share/initramfs-tools/hooks/zdev', which is also 
in package s390-tools 2.6.0-0ubuntu2
  dpkg-deb: error: paste subprocess was killed by signal (Broken pipe)
  Errors were encountered while processing:
   /var/cache/apt/archives/zfsutils-linux_0.7.9-3ubuntu5_s390x.deb
  E: Sub-process /usr/bin/dpkg returned an error code (1)
  Exit request sent.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/s390-tools/+bug/1788314/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1788432] [NEW] 4.15 s390x kernel BUG at /build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!

2018-08-22 Thread Colin Ian King
Public bug reported:

uname -a
Linux ckingvm1 4.15.0-33-generic #36-Ubuntu SMP Wed Aug 15 13:42:17 UTC 2018 
s390x s390x s390x GNU/Linux

and same for 4.15.0-29-generic and 4.17.0-8-generic

./stress-ng --sysfs 0 -t 60

.. wait a few seconds and then:

[  119.445891] [ cut here ]
[  119.445898] kernel BUG at 
/build/linux-Gycr4Z/linux-4.15.0/drivers/block/virtio_blk.c:565!
[  119.446093] illegal operation: 0001 ilc:1 [#3] SMP 
[  119.446100] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic vfio_ccw sha512_s390 sha256_s390 vfio_mdev sha1_s390 sha_common 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi nfsd auth_rpcgss nfs_acl 
lockd grace sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate raid10 
raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
libcrc32c raid1 raid0 linear virtio_net crc32_vx_s390 virtio_blk
[  119.446166] CPU: 1 PID: 5420 Comm: stress-ng-sysfs Tainted: P  DO
 4.15.0-33-generic #36-Ubuntu
[  119.446168] Hardware name: IBM 2964 N63 400 (KVM/Linux)
[  119.446170] Krnl PSW : 12d313d3 405835bc 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
[  119.446177]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
[  119.446194] Krnl GPRS: de6dc5c2779af7d7 7ffaba20 0040 
6545
[  119.446196]03ff800058da 6546 6bf537c0 
6b60a100
[  119.446198] 00690648 7cc3de40 
7a74b000
[  119.446202]03ff80008210  03ff800058da 
7ac1bce8
[  119.446210] Krnl Code: 03ff80005912: ebbff0a80004lmg 
%r11,%r15,168(%r15)
[  119.446210]03ff80005918: c0f40560brcl
15,3ff800063d8
[  119.446210]   #03ff8000591e: a7f40001brc 
15,3ff80005920
[  119.446210]   >03ff80005922: 0707bcr 0,%r7
[  119.446210]03ff80005924: 0707bcr 0,%r7
[  119.446210]03ff80005926: 0707bcr 0,%r7
[  119.446210]03ff80005928: c004brcl
0,3ff80005928
[  119.446210]03ff8000592e: eb6ff0480024stmg
%r6,%r15,72(%r15)
[  119.446226] Call Trace:
[  119.446229] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
[  119.446234]  [<00690684>] dev_attr_show+0x3c/0x80 
[  119.446240]  [<00424ab4>] sysfs_kf_seq_show+0xbc/0x1a8 
[  119.446259]  [<003b048c>] seq_read+0xec/0x4c8 
[  119.446262]  [<003821ea>] vfs_read+0x8a/0x150 
[  119.446274]  [<00382786>] SyS_read+0x66/0xe0 
[  119.446278]  [<008e3028>] system_call+0xdc/0x2c8 
[  119.446279] Last Breaking-Event-Address:
[  119.446281]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
[  119.446283]  
[  119.446284] ---[ end trace 2c2403d726047e4a ]---


For  4.17.0-8-generic:
[   25.170715] kernel BUG at drivers/block/virtio_blk.c:574!
[   25.170795] illegal operation: 0001 ilc:1 [#1] SMP 
[   25.170797] Modules linked in: lttng_statedump(OE) lttng_clock(OE) 
lttng_lib_ring_buffer(OE) binfmt_misc zfs(PO) zunicode(PO) zavl(PO) icp(PO) 
isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm nfsd 
ib_core auth_rpcgss iscsi_tcp nfs_acl lockd grace libiscsi_tcp libiscsi 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear virtio_net virtio_blk crc32_vx_s390
[   25.170835] CPU: 0 PID: 5590 Comm: stress-ng-sysfs Tainted: P   OE   
  4.17.0-8-generic #9-Ubuntu
[   25.170837] Hardware name: IBM 2964 N63 400 (KVM/Linux)
[   25.170839] Krnl PSW : 05f0c968 26542d57 
(virtblk_cache_type_show+0x7c/0x80 [virtio_blk])
[   25.170846]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
[   25.170849] Krnl GPRS: de6dc5c209bf9e6f 7ff91320 0040 
0891
[   25.170850] 0892  
7176e800
[   25.170852]6f447f00 00684300 6f492b40 
6ca7a000
[   25.170853]03ff80018218 7b5e6e40 03ff8001561a 
6f447ce8
[   25.170861] Krnl Code: 03ff8001564c: ebbff0a80004lmg 
%r11,%r15,168(%r15)
[   25.170861]03ff80015652: c0f4065bbrcl
15,3ff80016308
[   25.170861]   #03ff80015658: a7f40001brc 

[Kernel-packages] [Bug 1788385] Re: vfork in ubuntu_stress_smoke test failed on KVM node with B-KVM kernel

2018-08-22 Thread Colin Ian King
Do you mind providing information on the architecture and kernel version
+ release that is being tested.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1788385

Title:
  vfork in ubuntu_stress_smoke test failed on KVM node with B-KVM kernel

Status in Stress-ng:
  New
Status in ubuntu-kernel-tests:
  New
Status in linux-kvm package in Ubuntu:
  New

Bug description:
  Fail rate 2 out of 2, failed on a kvm node "gummo", with 1G ram

  DEBUG - 11:24:24 DEBUG| [stdout] vfork STARTING
  DEBUG - 11:24:24 DEBUG| [stdout] vfork STARTING
  DEBUG - 11:24:26 DEBUG| [stdout] vfork RETURNED 2
  DEBUG - 11:24:26 DEBUG| [stdout] vfork RETURNED 2
  DEBUG - 11:24:26 DEBUG| [stdout] vfork FAILED
  DEBUG - 11:24:26 DEBUG| [stdout] vfork FAILED
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 2 
processors online, 2 processors configured
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 2 
processors online, 2 processors configured
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: info:  [1437] 
dispatching hogs: 4 vfork
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: info:  [1437] 
dispatching hogs: 4 vfork
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: reducing cache level from L3 (too high) to L2
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: reducing cache level from L3 (too high) to L2
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: default cache size: 4096K
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] cache 
allocate: default cache size: 4096K
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 
starting stressors
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 
starting stressors
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1438] 
stress-ng-vfork: started [1438] (instance 0)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1438] 
stress-ng-vfork: started [1438] (instance 0)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 4 
stressors spawned
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1437] 4 
stressors spawned
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1439] 
stress-ng-vfork: started [1439] (instance 1)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1439] 
stress-ng-vfork: started [1439] (instance 1)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1440] 
stress-ng-vfork: started [1440] (instance 2)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1440] 
stress-ng-vfork: started [1440] (instance 2)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1441] 
stress-ng-vfork: started [1441] (instance 3)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: debug: [1441] 
stress-ng-vfork: started [1441] (instance 3)
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1438] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1438] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1440] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork failed
  DEBUG - 11:24:26 DEBUG| [stdout] stress-ng: fail:  [1441] 
stress-ng-vfork: fork 

[Kernel-packages] [Bug 1759848] Re: Allow multiple mounts of zfs datasets

2018-08-21 Thread Colin Ian King
Tested with zfsutils-linux 0.6.5.6-0ubuntu24 and linux 4.4.0-134 with
test in comment #2 and this now works. Marking as verified for Xenial

** Tags removed: verification-needed-xenial
** Tags added: verification-done-xenial

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1759848

Title:
  Allow multiple mounts of zfs datasets

Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  Fix Committed
Status in linux source package in Bionic:
  Fix Released
Status in zfs-linux source package in Bionic:
  Fix Released

Bug description:
  === SRU Justification, Xenial ==

  An attempt to mount an already mounted zfs dataset should return a new
  mount referencing the existing super block, but instead it returns an
  error. Operations such as bind mounts and unsharing mount namespaces
  create new mounts for the sb, which can cause operations to fail which
  involve unmounting and remounting the dataset.

  == Fix ==

  Backport of upstream fix https://trello.com/c/l89Ygj28/352-allow-
  multiple-mounts-of-zfs-datasets to allow multiple mounts

  This fix from Seth addresses this issue

  == Regression potential ==

  Like all backports, this has a potential to be incorrectly backported
  and break the ZFS mounting. However, any breakage should be picked up
  via the ZFS smoke tests that thoroughly exercise mounting/dismounting
  options.  At worst, the mounting won't work, but this has been tested,
  so I doubt this is a possibility.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1759848/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1786462] Re: linux 4.17.0-7.8 ADT test failure with linux 4.17.0-7.8

2018-08-21 Thread Colin Ian King
Any update? Does this fix the issue?

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1786462

Title:
  linux 4.17.0-7.8 ADT test failure with linux 4.17.0-7.8

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete
Status in linux source package in Cosmic:
  Incomplete

Bug description:
  Testing failed on:
  i386: 
https://objectstorage.prodstack4-5.canonical.com/v1/AUTH_77e2ada1e7a84929a74ba3b87153c0ac/autopkgtest-cosmic/cosmic/i386/l/linux/20180810_012246_01e82@/log.gz

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1786462/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-08-19 Thread Colin Ian King
The bug will be fixed once the zfs package and the bionic kernel (that
contains the zfs driver changes) will be released. So far, just the zfs
package has been released and we are waiting for the kernel to complete
the SRU update and verification phase - this takes a bit longer as the
kernel contains a lot more other changes and we have to do more
exhaustive testing.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Committed
Status in zfs-linux source package in Bionic:
  Fix Released
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the 

[Kernel-packages] [Bug 1787309] Re: zed not sending mail on scrub finish

2018-08-16 Thread Colin Ian King
** Changed in: zfs-linux (Ubuntu)
   Importance: Undecided => High

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1787309

Title:
  zed not sending mail on scrub finish

Status in zfs-linux package in Ubuntu:
  New

Bug description:
  ZED should be sending email when scrub finishes but I do not get any email. I 
did update modify
  the zed.rc file to end the email by setting the following

  ZED_EMAIL_ADDR="root"
  ZED_EMAIL_PROG="mail"
  ZED_NOTIFY_VERBOSE=1

  
  I tested to verify mail work. I did the following:
  echo testing | mail -s testit root

  I did get the mail so the mail system work.

  I also check syslog to see verify that zed saw the event.

  Aug 15 22:45:56 mike-think systemd[1]: Started ZFS Event Daemon (zed).
  Aug 15 22:45:56 mike-think zed[5576]: ZFS Event Daemon 0.6.5.6-0ubuntu23 (PID 
5576)
  Aug 15 22:45:56 mike-think zed[5576]: Processing events since eid=7
  Aug 15 22:48:38 mike-think zed: eid=8 class=history_event pool=lxc
  Aug 15 22:48:38 mike-think zed: eid=9 class=scrub_finish pool=lxc

  mike@mike-think:/var/log$ lsb_release -rd
  Description:  Ubuntu 16.04.5 LTS
  Release:  16.04
  mike@mike-think:/var/log$ 

  mike@mike-think:/var/log$ apt-cache policy zfs-zed
  zfs-zed:
Installed: 0.6.5.6-0ubuntu23
Candidate: 0.6.5.6-0ubuntu23
Version table:
   *** 0.6.5.6-0ubuntu23 500
  500 http://us.archive.ubuntu.com/ubuntu xenial-updates/main amd64 
Packages
  100 /var/lib/dpkg/status
   0.6.5.6-0ubuntu8 500
  500 http://us.archive.ubuntu.com/ubuntu xenial/universe amd64 Packages
  mike@mike-think:/var/log$

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: zfs-zed 0.6.5.6-0ubuntu23
  ProcVersionSignature: Ubuntu 4.15.0-32.35~16.04.1-generic 4.15.18
  Uname: Linux 4.15.0-32-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl icp zcommon znvpair
  ApportVersion: 2.20.1-0ubuntu2.18
  Architecture: amd64
  Date: Wed Aug 15 22:50:27 2018
  InstallationDate: Installed on 2016-07-24 (752 days ago)
  InstallationMedia: Ubuntu-GNOME 16.04.1 LTS "Xenial Xerus" - Release amd64 
(20160720)
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  SourcePackage: zfs-linux
  UpgradeStatus: No upgrade log present (probably fresh install)
  mtime.conffile..etc.zfs.zed.d.zed.rc: 2018-08-15T22:44:06.422894

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1787309/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1729878] Re: stress-ng triggering oomkiller running brk and stack stressors on all arches.

2018-08-15 Thread Colin Ian King
These stressors are expected to trigger the OOM killer.  The oom
adjustment for the stressor child processes is always adjusted to make
them the first processes to be OOM'd.  However, the kernel make's it
choice on what gets OOM'd depending on many factors, so the stressors
may end up triggering other hoggy processes to be OOM'd as well.

This is not a bug, it is expected behaviour.  See the stress-ng manual,
it clearly states when a stressors can trigger the OOM killer and how
each stress-ng stressor copes with this, for example:

   --brk N
  start N workers that grow the data segment by one page at a time 
using
  multiple brk(2) calls. Each successfully allocated new page is 
touched
  to ensure it is resident in memory.  If an  out  of  memory  
condition
  occurs  then  the test will reset the data segment to the point 
before
  it started and repeat the  data  segment  resizing  over  again.  
 The
  process  adjusts the out of memory setting so that it may be 
killed by
  the out of memory (OOM) killer  before  other  processes.   If  
it  is
  killed by the OOM killer then it will be automatically re-started 
by a
  monitoring parent process.

I shall close this as it is not a bug. Stress-ng is expected to stress a
kernel to trigger OOMs.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1729878

Title:
  stress-ng triggering oomkiller running brk and stack stressors on all
  arches.

Status in Stress-ng:
  New
Status in linux package in Ubuntu:
  Confirmed
Status in linux source package in Artful:
  Won't Fix

Bug description:
  During regression testing on a z/VM instance, I noticed call traces
  being dumped to dmesg that seem to be related to stress-ng.

  The stress-ng invocation we're using is:
  stress-ng --aggressive --verify --timeout $runtime \
   --metrics-brief --tz --times \
   --af-alg 0 --bsearch 0 --context 0 --cpu 0 \
   --crypt 0 --hsearch 0 --longjmp 0 --lsearch 
0 \
   --matrix 0 --qsort 0 --str 0 --stream 0 \
   --tsearch 0 --vecmath 0 --wcs 0

  as executed by the cpu_stress script from the cpu/cpu_stress_ng test
  in the certification suite.

  ProblemType: Bug
  DistroRelease: Ubuntu 17.10
  Package: linux-image-4.13.0-16-generic 4.13.0-16.19
  ProcVersionSignature: Ubuntu 4.13.0-16.19-generic 4.13.4
  Uname: Linux 4.13.0-16-generic s390x
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory: 'aplay': 'aplay'
  ApportVersion: 2.20.7-0ubuntu3.1
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory: 'arecord': 
'arecord'
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  Date: Fri Nov  3 11:12:04 2017
  HibernationDevice: RESUME=UUID=1ce0f037-449a-43a1-af49-e730f6f99ac4
  IwConfig: Error: [Errno 2] No such file or directory: 'iwconfig': 'iwconfig'
  Lspci:
   
  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  PciMultimedia:
   
  ProcEnviron:
   TERM=screen
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=C
   SHELL=/bin/bash
  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=/dev/mapper/hwe0008_vg-lv crashkernel=196M 
BOOT_IMAGE=0
  RelatedPackageVersions:
   linux-restricted-modules-4.13.0-16-generic N/A
   linux-backports-modules-4.13.0-16-generic  N/A
   linux-firmware 1.169
  RfKill: Error: [Errno 2] No such file or directory: 'rfkill': 'rfkill'
  SourcePackage: linux
  UpgradeStatus: Upgraded to artful on 2017-11-02 (0 days ago)
  --- 
  AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 
2: ls: cannot access '/dev/snd/': No such file or directory
  AplayDevices: Error: [Errno 2] No such file or directory
  ApportVersion: 2.20.7-0ubuntu3.1
  Architecture: s390x
  ArecordDevices: Error: [Errno 2] No such file or directory
  CRDA: Error: command ['iw', 'reg', 'get'] failed with exit code 1: nl80211 
not found.
  DistroRelease: Ubuntu 17.10
  HibernationDevice: RESUME=UUID=1ce0f037-449a-43a1-af49-e730f6f99ac4
  IwConfig: Error: [Errno 2] No such file or directory
  Lspci:
   
  Lsusb: Error: command ['lsusb'] failed with exit code 1:
  Package: linux (not installed)
  PciMultimedia:
   
  ProcEnviron:
   TERM=screen
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=C
   SHELL=/bin/bash
  ProcFB: Error: [Errno 2] No such file or directory: '/proc/fb'
  ProcKernelCmdLine: root=/dev/mapper/hwe0008_vg-lv crashkernel=196M 
BOOT_IMAGE=0
  ProcVersionSignature: Ubuntu 

[Kernel-packages] [Bug 1786462] Re: linux 4.17.0-7.8 ADT test failure with linux 4.17.0-7.8

2018-08-14 Thread Colin Ian King
Potential fix applied: http://kernel.ubuntu.com/git/ubuntu/autotest-
client-tests.git/commit/?id=cf7f9e75bb1fe0f7d798a8d49e0407b602017848

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1786462

Title:
  linux 4.17.0-7.8 ADT test failure with linux 4.17.0-7.8

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete
Status in linux source package in Cosmic:
  Incomplete

Bug description:
  Testing failed on:
  i386: 
https://objectstorage.prodstack4-5.canonical.com/v1/AUTH_77e2ada1e7a84929a74ba3b87153c0ac/autopkgtest-cosmic/cosmic/i386/l/linux/20180810_012246_01e82@/log.gz

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1786462/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1786462] Re: linux 4.17.0-7.8 ADT test failure with linux 4.17.0-7.8

2018-08-10 Thread Colin Ian King
When running this test on 3 concurrent VMs with this kernel this runs fine.  
With the ZFS backing store of the VMs set to sync on writes, I can slow the I/O 
and I get a lot of timeout issues which result in tests running for a long time 
and probably would get nuked by the testing infrastructure - 
see attached dmesg log for the timeouts.

I am currently speculating that the tests are failing because VFAT is
sub-optimal compared to other file systems and with a slow backing store
we may be just getting lockup warnings because of slow I/O write out
and/or syncs.

At least I have a reproducer. I don't think this is a critical show
stopping bug. I will try and figure out what particular stress case is
causing the issue, but this may take a while to iterate through.


** Attachment added: "dmesg log"
   
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1786462/+attachment/5173892/+files/dmesg.log

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1786462

Title:
  linux 4.17.0-7.8 ADT test failure with linux 4.17.0-7.8

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete
Status in linux source package in Cosmic:
  Incomplete

Bug description:
  Testing failed on:
  i386: 
https://objectstorage.prodstack4-5.canonical.com/v1/AUTH_77e2ada1e7a84929a74ba3b87153c0ac/autopkgtest-cosmic/cosmic/i386/l/linux/20180810_012246_01e82@/log.gz

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1786462/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1785816] Re: Consider enabling CONFIG_NETWORK_PHY_TIMESTAMPING

2018-08-08 Thread Colin Ian King
@Seth, would you be so kind to as to enable this given the evidence
above?

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1785816

Title:
  Consider enabling CONFIG_NETWORK_PHY_TIMESTAMPING

Status in linux package in Ubuntu:
  Triaged

Bug description:
  Hi folks, is there a reason we do not enable
  CONFIG_NETWORK_PHY_TIMESTAMPING ? I'm not sure of the tradeoffs but I
  think this config option is required for PTP time syncronization which
  would be useful in telco and financial (and OCD :)) environments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1785816/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1785816] Re: Consider enabling CONFIG_NETWORK_PHY_TIMESTAMPING

2018-08-08 Thread Colin Ian King
With config CONFIG_NETWORK_PHY_TIMESTAMPING enabled, the calls to
skb_clone_tx_timestamp() and skb_defer_rx_timestamp() are enabled (these
normally are empty inlined no-op functions).  The overhead from what I
can see is very small, for example for the tx path:

static unsigned int classify(const struct sk_buff *skb)
{
if (likely(skb->dev && skb->dev->phydev &&
   skb->dev->phydev->drv))
return ptp_classify_raw(skb);  
else
return PTP_CLASS_NONE;
}

void skb_clone_tx_timestamp(struct sk_buff *skb)
{
struct phy_device *phydev;
struct sk_buff *clone;
unsigned int type;

if (!skb->sk)
return;

type = classify(skb);
if (type == PTP_CLASS_NONE)
return;

phydev = skb->dev->phydev;
if (likely(phydev->drv->txtstamp)) {
clone = skb_clone_sk(skb);
if (!clone)
return;
phydev->drv->txtstamp(phydev, clone, type);
}
}

The classify() call is an overhead that runs a minimal BPF dissector to 
classify a network packet to
determine the PTP class. For the default non PTP case this returns 
PTP_CLASS_NONE. The BPF classifier is just 3-4 BPF branches (depending on the 
protocol), so it's a very small overhead per packet in the default non-PTP 
cases.

I ran some perf timings on TCP data being sent and received to a host
over a 100 Mbit/s ethernet between to 8 thread Xeon servers and measured
CPU cycles, instruction and branch activity with perf.  1 GB of raw data
was transferred to/from the machines using netcat on otherwises idle
systems. Each test was run 10 times and the average, standard deviation
(population) and % standard deviation was computed.

I compared a default 4.17.0-6-generic Ubuntu Cosmic kernel against the
same kernel with CONFIG_NETWORK_PHY_TIMESTAMPING.  I could not observe
any noticeable impact with the CONFIG_NETWORK_PHY_TIMESTAMPING config -
mainly because the noise in the perf measurements was larger than any
detectable difference (see the % standard deviation rates).

Since I can't easily measure the performance impact any more accurately
than instruction and branch counts, I conclude that the impact of this
config is not easily measurable and too small to be a concern.

Data in a libreoffice spread sheet is attached.

I therefore deem this config is OK to be enabled for by default for our
kernels.

** Attachment added: "libreoffice spreadsheet with test results data"
   
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1785816/+attachment/5173042/+files/kernel-config-CONFIG_NETWORK_PHY_TIMESTAMPING.ods

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1785816

Title:
  Consider enabling CONFIG_NETWORK_PHY_TIMESTAMPING

Status in linux package in Ubuntu:
  Triaged

Bug description:
  Hi folks, is there a reason we do not enable
  CONFIG_NETWORK_PHY_TIMESTAMPING ? I'm not sure of the tradeoffs but I
  think this config option is required for PTP time syncronization which
  would be useful in telco and financial (and OCD :)) environments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1785816/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-08-07 Thread Colin Ian King
@Vasiliy, hopefully by early next week.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Committed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 1 >/sys/module/spl/parameters/spl_taskq_kick

  but did not manage to continue.
  Include any warning/errors/backtraces from the system logs
  

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-08-07 Thread Colin Ian King
Verified passed for Ubuntu Bionic using the reproducer described in
comment #1.  Marking as verified.

** Tags removed: verification-needed-bionic
** Tags added: verification-done-bionic

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Committed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-08-07 Thread Colin Ian King
Verified passed for Ubuntu Xenial using the reproducer described in
comment #1. Marking as verified.

** Tags removed: verification-needed-xenial
** Tags added: verification-done-xenial

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Released
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Fix Committed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Fix Committed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Released
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 

[Kernel-packages] [Bug 1785816] Re: Consider enabling CONFIG_NETWORK_PHY_TIMESTAMPING

2018-08-07 Thread Colin Ian King
I'll rig up some test scenarios and measure the overhead.

** Changed in: linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1785816

Title:
  Consider enabling CONFIG_NETWORK_PHY_TIMESTAMPING

Status in linux package in Ubuntu:
  Triaged

Bug description:
  Hi folks, is there a reason we do not enable
  CONFIG_NETWORK_PHY_TIMESTAMPING ? I'm not sure of the tradeoffs but I
  think this config option is required for PTP time syncronization which
  would be useful in telco and financial (and OCD :)) environments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1785816/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1772617] Re: fix inverted boolean flag in arch_add_memory, reverts back to original behaviour

2018-08-06 Thread Colin Ian King
Artful is now EOL, closing bug

** Changed in: linux (Ubuntu)
   Status: In Progress => Won't Fix

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1772617

Title:
  fix inverted boolean flag in arch_add_memory, reverts back to original
  behaviour

Status in linux package in Ubuntu:
  Won't Fix

Bug description:
  == SRU Justification, ARTFUL ==

  Bug fix #1761104 incorrectly inverted the flag in the call to 
arch_add_memory, it should be true instead of false.
  == Fix ==

  Fix partial backport from bug #1747069, remove can_online_high_movable
  and fix the incorrectly set boolean argument to arch_add_memory().
  NVIDIA are reported that this is was incorrectly flipped in the last
  SRU to fix their driver support on powerpc.

  == Testing ==

  run ADT memory hotplug test, should not regress this. Without the fix,
  the nvidia driver on powerpc will not work. With the fix it loads and
  works.

  == Regression Potential ==

  This fixes a regression in the original fix and hence the regression
  potential is the same as the previously SRU'd bug fix for #1747069,
  namely:

  "Reverting this commit does remove some functionality, however this
  does not regress the kernel compared to previous releases and having a
  working reliable memory hotplug is the preferred option. This fix does
  touch some memory hotplug, so there is a risk that this may break this
  functionality that is not covered by the kernel regression testing."

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1772617/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1783651] Re: Please enable CONFIG_PAGE_POISONING

2018-08-02 Thread Colin Ian King
I've built and tested 4.15 and the latest 4.18 with the following
configs:

CONFIG_PAGE_POISONING=y
CONFIG_PAGE_POISONING_ZERO=y
CONFIG_PAGE_POISONING_NO_SANITY=y

Attached is a libreoffice spreadsheet with the test results comparing
kernels without the config, with the config and with the config with
page_poison=1 for the 4.18 and 4.15 kernels. I ran nearly 200 stress-ng
stress tests and gathered the throughput (based on bogo ops per second
on the usr+sys time consumed) for each stress test.  Each of the stress
tests were run for 60 seconds on an idle 8 thread Xeon i7-3770.

The bogo-ops data was then normalized against the kernel that didn't
have the config changes.   The data to look at is the geometric means of
all the normalized test results:

4.18 kernel, geometric mean of normalized bogo/ops throughput:

No page poisoning:  1.000
Config page poisoning: 1.003
Config page poionsing + page_poison=1: 0.991

4.15 kernel, geometric mean of normalized bogo/ops throughput:

No page poisoning:  1.000
Config page poisoning: 1.025
Config page poionsing + page_poison=1: 0.977

where > 1.000 shows more throughput and < 1.000 shows degraded
throughput

So it appears that enabling page poisoning configs does not degrade
performance and setting page_poison=1 degrades performance by a very
small amount.

** Attachment added: "libreoffice spreadsheet with test results data"
   
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1783651/+attachment/5170997/+files/kernel-poison-page-analysis.ods

** Changed in: linux (Ubuntu Cosmic)
 Assignee: Colin Ian King (colin-king) => (unassigned)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1783651

Title:
  Please enable CONFIG_PAGE_POISONING

Status in linux package in Ubuntu:
  Confirmed
Status in linux source package in Cosmic:
  Confirmed

Bug description:
  I'd like to be able to use page poisoning, but CONFIG_PAGE_POISONING
  is not enabled on Ubuntu. (This option itself has a near-zero
  performance impact since it must be combined with the boot option
  "page_poison=1" to actually enable the poisoning.)

  To make the poisoning (when enabled) less of an impact, I'd also like
  to see CONFIG_PAGE_POISONING_ZERO=y too, which means GFP_ZEROing can
  be skipped.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1783651/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1783651] Re: Please enable CONFIG_PAGE_POISONING

2018-08-01 Thread Colin Ian King
** Changed in: linux (Ubuntu Cosmic)
 Assignee: (unassigned) => Colin Ian King (colin-king)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1783651

Title:
  Please enable CONFIG_PAGE_POISONING

Status in linux package in Ubuntu:
  Confirmed
Status in linux source package in Cosmic:
  Confirmed

Bug description:
  I'd like to be able to use page poisoning, but CONFIG_PAGE_POISONING
  is not enabled on Ubuntu. (This option itself has a near-zero
  performance impact since it must be combined with the boot option
  "page_poison=1" to actually enable the poisoning.)

  To make the poisoning (when enabled) less of an impact, I'd also like
  to see CONFIG_PAGE_POISONING_ZERO=y too, which means GFP_ZEROing can
  be skipped.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1783651/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1760637] Re: af-alg test in ubuntu_stress_smoke_test failed with 4.4/4.15 KVM kernel

2018-07-20 Thread Colin Ian King
I've also added changes to stress-ng to warn this protocol is not
supported rather than fail:

http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=3ec28f2f54381910771299df03c0dfa90a3bdbf5

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1760637

Title:
  af-alg test in ubuntu_stress_smoke_test failed with 4.4/4.15 KVM
  kernel

Status in ubuntu-kernel-tests:
  Fix Released
Status in linux package in Ubuntu:
  Invalid
Status in linux-kvm package in Ubuntu:
  Invalid
Status in stress-ng package in Ubuntu:
  Fix Released

Bug description:
  The test failed with:
  16:23:49 DEBUG| [stdout] af-alg STARTING
  16:23:54 DEBUG| [stdout] af-alg RETURNED 2
  16:23:54 DEBUG| [stdout] af-alg FAILED
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] 2 processors online, 2 
processors configured
  16:23:54 DEBUG| [stdout] stress-ng: info:  [3569] dispatching hogs: 4 af-alg
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] cache allocate: reducing 
cache level from L3 (too high) to L2
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] cache allocate: default 
cache size: 4096K
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] starting stressors
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] 4 stressors spawned
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3570] stress-ng-af-alg: started 
[3570] (instance 0)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3573] stress-ng-af-alg: started 
[3573] (instance 3)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3571] stress-ng-af-alg: started 
[3571] (instance 1)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3572] stress-ng-af-alg: started 
[3572] (instance 2)
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3570] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3570] stress-ng-af-alg: exited 
[3570] (instance 0)
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3570 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3570] terminated
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3571] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3571] stress-ng-af-alg: exited 
[3571] (instance 1)
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3571 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3571] terminated
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3573] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3573] stress-ng-af-alg: exited 
[3573] (instance 3)
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3572] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3572] stress-ng-af-alg: exited 
[3572] (instance 2)
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3572 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3572] terminated
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3573 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3573] terminated
  16:23:54 DEBUG| [stdout] stress-ng: info:  [3569] unsuccessful run completed 
in 5.01s
  16:23:54 DEBUG| [stdout]

  Steps to reproduce:
    Deploy the node with Xenial 4.4 kernel, install linux-kvm
    sudo apt-get install python-minimal
    git clone --depth=1 git://kernel.ubuntu.com/ubuntu/autotest-client-tests -b 
master-next
    git clone --depth=1 git://kernel.ubuntu.com/ubuntu/autotest
    rm -fr autotest/client/tests
    ln -sf ~/autotest-client-tests autotest/client/tests
    AUTOTEST_PATH=/home/ubuntu/autotest sudo -E autotest/client/autotest-local 
--verbose autotest/client/tests/ubuntu_stress_smoke_test/control 

  
  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-1019-kvm 4.4.0-1019.24
  ProcVersionSignature: User Name 4.4.0-1019.24-kvm 4.4.98
  Uname: Linux 4.4.0-1019-kvm x86_64
  ApportVersion: 2.20.1-0ubuntu2.15
  Architecture: amd64
  Date: Mon Apr  2 16:25:57 2018
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  SourcePackage: linux-kvm
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1760637/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : 

[Kernel-packages] [Bug 1760637] Re: af-alg test in ubuntu_stress_smoke_test failed with 4.4/4.15 KVM kernel

2018-07-20 Thread Colin Ian King
This kernel feature is disabled, see bug:
https://bugs.launchpad.net/ubuntu/+source/linux-kvm/+bug/1782728

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1760637

Title:
  af-alg test in ubuntu_stress_smoke_test failed with 4.4/4.15 KVM
  kernel

Status in ubuntu-kernel-tests:
  Fix Released
Status in linux package in Ubuntu:
  Invalid
Status in linux-kvm package in Ubuntu:
  Invalid
Status in stress-ng package in Ubuntu:
  Fix Released

Bug description:
  The test failed with:
  16:23:49 DEBUG| [stdout] af-alg STARTING
  16:23:54 DEBUG| [stdout] af-alg RETURNED 2
  16:23:54 DEBUG| [stdout] af-alg FAILED
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] 2 processors online, 2 
processors configured
  16:23:54 DEBUG| [stdout] stress-ng: info:  [3569] dispatching hogs: 4 af-alg
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] cache allocate: reducing 
cache level from L3 (too high) to L2
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] cache allocate: default 
cache size: 4096K
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] starting stressors
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] 4 stressors spawned
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3570] stress-ng-af-alg: started 
[3570] (instance 0)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3573] stress-ng-af-alg: started 
[3573] (instance 3)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3571] stress-ng-af-alg: started 
[3571] (instance 1)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3572] stress-ng-af-alg: started 
[3572] (instance 2)
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3570] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3570] stress-ng-af-alg: exited 
[3570] (instance 0)
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3570 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3570] terminated
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3571] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3571] stress-ng-af-alg: exited 
[3571] (instance 1)
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3571 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3571] terminated
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3573] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3573] stress-ng-af-alg: exited 
[3573] (instance 3)
  16:23:54 DEBUG| [stdout] stress-ng: fail:  [3572] stress-ng-af-alg: socket 
failed, errno=97 (Address family not supported by protocol)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3572] stress-ng-af-alg: exited 
[3572] (instance 2)
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3572 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3572] terminated
  16:23:54 DEBUG| [stdout] stress-ng: error: [3569] process 3573 
(stress-ng-af_alg) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:23:54 DEBUG| [stdout] stress-ng: debug: [3569] process [3573] terminated
  16:23:54 DEBUG| [stdout] stress-ng: info:  [3569] unsuccessful run completed 
in 5.01s
  16:23:54 DEBUG| [stdout]

  Steps to reproduce:
    Deploy the node with Xenial 4.4 kernel, install linux-kvm
    sudo apt-get install python-minimal
    git clone --depth=1 git://kernel.ubuntu.com/ubuntu/autotest-client-tests -b 
master-next
    git clone --depth=1 git://kernel.ubuntu.com/ubuntu/autotest
    rm -fr autotest/client/tests
    ln -sf ~/autotest-client-tests autotest/client/tests
    AUTOTEST_PATH=/home/ubuntu/autotest sudo -E autotest/client/autotest-local 
--verbose autotest/client/tests/ubuntu_stress_smoke_test/control 

  
  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: linux-image-4.4.0-1019-kvm 4.4.0-1019.24
  ProcVersionSignature: User Name 4.4.0-1019.24-kvm 4.4.98
  Uname: Linux 4.4.0-1019-kvm x86_64
  ApportVersion: 2.20.1-0ubuntu2.15
  Architecture: amd64
  Date: Mon Apr  2 16:25:57 2018
  ProcEnviron:
   TERM=xterm-256color
   PATH=(custom, no user)
   XDG_RUNTIME_DIR=
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  SourcePackage: linux-kvm
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1760637/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1760638] Re: dccp test in ubuntu_stress_smoke_test failed with 4.4/4.15 KVM kernel

2018-07-20 Thread Colin Ian King
See also: https://bugs.launchpad.net/ubuntu/+source/linux-
kvm/+bug/1782727

DCCP is not enabled in this kernel

..and updates stress-ng to warn and not fail if the protocol is
disabled:

http://kernel.ubuntu.com/git/cking/stress-
ng.git/commit/?id=1c54d53b33d8e2b11c9c47b9cff04612ba9d7e14

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1760638

Title:
  dccp test in ubuntu_stress_smoke_test failed with 4.4/4.15 KVM kernel

Status in ubuntu-kernel-tests:
  Fix Released
Status in linux package in Ubuntu:
  Invalid
Status in linux-kvm package in Ubuntu:
  Invalid
Status in stress-ng package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Invalid
Status in linux-kvm source package in Xenial:
  Invalid
Status in stress-ng source package in Xenial:
  Fix Released
Status in linux source package in Bionic:
  Invalid
Status in linux-kvm source package in Bionic:
  Invalid
Status in stress-ng source package in Bionic:
  Fix Released

Bug description:
  The test failed with:
  16:26:24 DEBUG| [stdout] dccp STARTING
  16:26:24 DEBUG| [stdout] dccp RETURNED 2
  16:26:24 DEBUG| [stdout] dccp FAILED
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] 2 processors online, 2 
processors configured
  16:26:24 DEBUG| [stdout] stress-ng: info:  [2228] dispatching hogs: 4 dccp
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] cache allocate: reducing 
cache level from L3 (too high) to L2
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] cache allocate: default 
cache size: 4096K
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] starting stressors
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] 4 stressors spawned
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2231] stress-ng-dccp: started 
[2231] (instance 2)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2232] stress-ng-dccp: started 
[2232] (instance 3)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2229] stress-ng-dccp: started 
[2229] (instance 0)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2230] stress-ng-dccp: started 
[2230] (instance 1)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2231] stress-ng-dccp: process 
[2231] using socket port 10002
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2229] stress-ng-dccp: process 
[2229] using socket port 1
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2232] stress-ng-dccp: process 
[2232] using socket port 10003
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2230] stress-ng-dccp: process 
[2230] using socket port 10001
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2231] stress-ng-dccp: socket 
failed, errno=94 (Socket type not supported)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2231] stress-ng-dccp: 0 messages 
sent
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2231] stress-ng-dccp: exited 
[2231] (instance 2)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2239] stress-ng-dccp: socket 
failed, errno=94 (Socket type not supported)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2230] stress-ng-dccp: socket 
failed, errno=94 (Socket type not supported)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2230] stress-ng-dccp: 0 messages 
sent
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2230] stress-ng-dccp: exited 
[2230] (instance 1)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2232] stress-ng-dccp: socket 
failed, errno=94 (Socket type not supported)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2232] stress-ng-dccp: 0 messages 
sent
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2232] stress-ng-dccp: exited 
[2232] (instance 3)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2235] stress-ng-dccp: socket 
failed, errno=94 (Socket type not supported)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2229] stress-ng-dccp: socket 
failed, errno=94 (Socket type not supported)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2229] stress-ng-dccp: 0 messages 
sent
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2229] stress-ng-dccp: exited 
[2229] (instance 0)
  16:26:24 DEBUG| [stdout] stress-ng: error: [2228] process 2229 
(stress-ng-dccp) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] process [2229] terminated
  16:26:24 DEBUG| [stdout] stress-ng: error: [2228] process 2230 
(stress-ng-dccp) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] process [2230] terminated
  16:26:24 DEBUG| [stdout] stress-ng: error: [2228] process 2231 
(stress-ng-dccp) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] process [2231] terminated
  16:26:24 DEBUG| [stdout] stress-ng: error: [2228] process 2232 
(stress-ng-dccp) terminated with an error, exit status=1 (stress-ng core 
failure)
  16:26:24 DEBUG| [stdout] stress-ng: debug: [2228] process [2232] terminated
  16:26:24 DEBUG| [stdout] stress-ng: info:  [2228] 

[Kernel-packages] [Bug 1782728] [NEW] crypto AF_ALG protocol not enabled on KVM - is this intentional?

2018-07-20 Thread Colin Ian King
Public bug reported:

The 4.15.0 KVM kernels do not enable AF_ALG crypto [1] by default.  This
userspace crypto can be useful - so is it intentional for this to be
disabled?

[1] https://en.wikipedia.org/wiki/Crypto_API_(Linux)

** Affects: linux-kvm (Ubuntu)
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1782728

Title:
  crypto AF_ALG protocol not enabled on KVM - is this intentional?

Status in linux-kvm package in Ubuntu:
  New

Bug description:
  The 4.15.0 KVM kernels do not enable AF_ALG crypto [1] by default.
  This userspace crypto can be useful - so is it intentional for this to
  be disabled?

  [1] https://en.wikipedia.org/wiki/Crypto_API_(Linux)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-kvm/+bug/1782728/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1782727] [NEW] DCCP protocol not enabled on KVM kernel - is this intentional?

2018-07-20 Thread Colin Ian King
Public bug reported:

kvm 4.15.0 kernels don't have DCCP [1] enabled by default unlike our
generic kernel.  Is this intentional?

[1] https://en.wikipedia.org/wiki/Datagram_Congestion_Control_Protocol

** Affects: linux-kvm (Ubuntu)
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1782727

Title:
  DCCP protocol not enabled on KVM kernel - is this intentional?

Status in linux-kvm package in Ubuntu:
  New

Bug description:
  kvm 4.15.0 kernels don't have DCCP [1] enabled by default unlike our
  generic kernel.  Is this intentional?

  [1] https://en.wikipedia.org/wiki/Datagram_Congestion_Control_Protocol

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-kvm/+bug/1782727/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1782723] Re: dnotify in ubuntu_stress_smoke_test failed on B-KVM

2018-07-20 Thread Colin Ian King
This also occurs in 4.15.0-1012-kvm so it is NOT a regression.

Config has:

# CONFIG_DNOTIFY is not set

so the failure is because DNOTIFY has be disables, which I believe is an
issue, this functionality should be enabled IMHO.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1782723

Title:
  dnotify in ubuntu_stress_smoke_test failed on B-KVM

Status in ubuntu-kernel-tests:
  New
Status in linux-kvm package in Ubuntu:
  New
Status in stress-ng package in Ubuntu:
  New

Bug description:
  This issue does not exist in 4.15.0-1015.15 - kvm

  
  08:08:02 DEBUG| [stdout] dnotify FAILED
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3043] 2 processors online, 2 
processors configured
  08:08:02 DEBUG| [stdout] stress-ng: info:  [3043] dispatching hogs: 4 dnotify
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3043] cache allocate: reducing 
cache level from L3 (too high) to L2
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3043] cache allocate: default 
cache size: 4096K
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3043] starting stressors
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3044] stress-ng-dnotify: started 
[3044] (instance 0)
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3043] 4 stressors spawned
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3044] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3046] stress-ng-dnotify: started 
[3046] (instance 2)
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3047] stress-ng-dnotify: started 
[3047] (instance 3)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3044] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3046] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3044] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3046] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3046] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3046] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3046] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] info: 5 failures reached, aborting stress process
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3046] stress-ng-dnotify: exited 
[3046] (instance 2)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3047] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3047] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3047] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3047] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3047] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] info: 5 failures reached, aborting stress process
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3047] stress-ng-dnotify: exited 
[3047] (instance 3)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3044] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3044] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] info: 5 failures reached, aborting stress process
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3044] stress-ng-dnotify: exited 
[3044] (instance 0)
  08:08:02 DEBUG| [stdout] stress-ng: error: [3043] process 3044 
(stress-ng-dnotify) terminated with an error, exit status=1 (stress-ng core 
failure)
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3043] process [3044] terminated
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3045] stress-ng-dnotify: started 
[3045] (instance 1)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3045] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3045] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3045] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3045] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] stress-ng: fail:  [3045] stress-ng-dnotify: fcntl 
F_NOTIFY failed, errno=22 (Invalid argument)
  08:08:02 DEBUG| [stdout] info: 5 failures reached, aborting stress process
  08:08:02 DEBUG| [stdout] stress-ng: debug: [3045] 

[Kernel-packages] [Bug 1782565] [NEW] s390 illegal operation

2018-07-19 Thread Colin Ian King
Public bug reported:

Managed to break sysfs when running:

sudo ./stress-ng  --sysfs 4 -t 60 -v --verify


Kernel: 4.15.0-29-generic and 4.17 too, this seems to occur on previous 4.15 
kernels but I've not tested too far back.


[   45.057702] [ cut here ]
[   45.057704] kernel BUG at 
/build/linux-4OqUyO/linux-4.15.0/drivers/block/virtio_blk.c:565!
[   45.057773] illegal operation: 0001 ilc:1 [#1] SMP 
[   45.057775] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi nfsd auth_rpcgss nfs_acl lockd grace 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear crc32_vx_s390 virtio_net virtio_blk
[   45.057805] CPU: 0 PID: 5056 Comm: stress-ng-sysfs Tainted: P   O
 4.15.0-29-generic #31-Ubuntu
[   45.057806] Hardware name: IBM 2964 N63 400 (KVM/Linux)
[   45.057809] Krnl PSW : 2a20609d 800ebb38 
(virtblk_cache_type_show+0x82/0x88 [virtio_blk])
[   45.057814]R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:2 PM:0 
RI:0 EA:3
[   45.057816] Krnl GPRS: de6dc5c2778317a7 7ff93a20 0080 
1164
[   45.057818]03ff800058da 1165 6b9cc840 
7061a100
[   45.057819] 0068feb8 72c30d80 
71972000
[   45.057821]03ff80008210 1000 03ff800058da 
70a47ce8
[   45.057830] Krnl Code: 03ff80005912: ebbff0a80004lmg 
%r11,%r15,168(%r15)
  03ff80005918: c0f40560brcl
15,3ff800063d8
 #03ff8000591e: a7f40001brc 
15,3ff80005920
 >03ff80005922: 0707bcr 0,%r7
  03ff80005924: 0707bcr 0,%r7
  03ff80005926: 0707bcr 0,%r7
  03ff80005928: c004brcl
0,3ff80005928
  03ff8000592e: eb6ff0480024stmg
%r6,%r15,72(%r15)
[   45.057844] Call Trace:
[   45.057847] ([<03ff800058da>] virtblk_cache_type_show+0x3a/0x88 
[virtio_blk])
[   45.057852]  [<0068fef4>] dev_attr_show+0x3c/0x80 
[   45.057858]  [<00424414>] sysfs_kf_seq_show+0xbc/0x1a8 
[   45.057862]  [<003afddc>] seq_read+0xec/0x4c8 
[   45.057865]  [<00381bba>] vfs_read+0x8a/0x150 
[   45.057867]  [<00382156>] SyS_read+0x66/0xe0 
[   45.057872]  [<008e2350>] system_call+0xdc/0x2c8 
[   45.057873] Last Breaking-Event-Address:
[   45.057875]  [<03ff8000591e>] virtblk_cache_type_show+0x7e/0x88 
[virtio_blk]
[   45.057876]  
[   45.057877] ---[ end trace 49fd8401e53e0e99 ]---

** Affects: linux (Ubuntu)
 Importance: Undecided
 Status: New

** Affects: linux (Ubuntu Bionic)
 Importance: Undecided
 Status: New

** Affects: linux (Ubuntu Cosmic)
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1782565

Title:
  s390 illegal operation

Status in linux package in Ubuntu:
  New
Status in linux source package in Bionic:
  New
Status in linux source package in Cosmic:
  New

Bug description:
  Managed to break sysfs when running:

  sudo ./stress-ng  --sysfs 4 -t 60 -v --verify

  
  Kernel: 4.15.0-29-generic and 4.17 too, this seems to occur on previous 4.15 
kernels but I've not tested too far back.

  
  [   45.057702] [ cut here ]
  [   45.057704] kernel BUG at 
/build/linux-4OqUyO/linux-4.15.0/drivers/block/virtio_blk.c:565!
  [   45.057773] illegal operation: 0001 ilc:1 [#1] SMP 
  [   45.057775] Modules linked in: binfmt_misc zfs(PO) zunicode(PO) zavl(PO) 
icp(PO) isofs zcommon(PO) znvpair(PO) spl(O) ghash_s390 prng aes_s390 des_s390 
des_generic sha512_s390 sha256_s390 sha1_s390 sha_common vfio_ccw vfio_mdev 
mdev vfio_iommu_type1 vfio sch_fq_codel ib_iser rdma_cm iw_cm ib_cm ib_core 
iscsi_tcp libiscsi_tcp libiscsi nfsd auth_rpcgss nfs_acl lockd grace 
scsi_transport_iscsi sunrpc ip_tables x_tables btrfs zstd_compress zlib_deflate 
raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor 
raid6_pq libcrc32c raid1 raid0 linear crc32_vx_s390 virtio_net virtio_blk
  [   45.057805] CPU: 0 PID: 5056 Comm: stress-ng-sysfs Tainted: P   O  
   4.15.0-29-generic #31-Ubuntu
  [   45.057806] Hardware name: IBM 2964 N63 400 (KVM/Linux)
  [   45.057809] Krnl PSW : 

[Kernel-packages] [Bug 1781364] Re: Kernel error "task zfs:pid blocked for more than 120 seconds"

2018-07-18 Thread Colin Ian King
The bug will be automatically updated when the -proposed kernel
containing the fix is ready, please wait for that message.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1781364

Title:
  Kernel error "task zfs:pid blocked for more than 120 seconds"

Status in Linux:
  Fix Released
Status in linux package in Ubuntu:
  Fix Committed
Status in zfs-linux package in Ubuntu:
  Fix Released
Status in linux source package in Xenial:
  Confirmed
Status in zfs-linux source package in Xenial:
  Confirmed
Status in linux source package in Bionic:
  Confirmed
Status in zfs-linux source package in Bionic:
  Fix Committed
Status in linux source package in Cosmic:
  Fix Committed
Status in zfs-linux source package in Cosmic:
  Fix Released

Bug description:
  == SRU Justification, XENIAL, BIONIC ==

  Exercising ZFS with lxd with many mount/umounts can cause lockups and
  120 second timeout messages.

  == How to reproduce bug ==

  In a VM, 2 CPUs, 16GB of memory running Bionic:

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux
  sudo lxd init

  (and with the default init options)

  then run:

  lxd-benchmark launch --count 96 --parallel 96

  This will reliably show the lockup every time without the fix.  With
  the fix (detailed below) one cannot reproduce the lockup.

  == Fix ==

  Upstream ZFS commit

  commit ac09630d8b0bf6c92084a30fdaefd03fd0adbdc1
  Author: Brian Behlendorf 
  Date: Wed Jul 11 15:49:10 2018 -0700

  Fix zpl_mount() deadlock

  == Regression Potential ==

  This just changes the locking in the mount path of ZFS and will only
  affect ZFS mount/unmounts.  The regression potential is small as this
  touches a very small code path that has been exhaustively exercises
  this code under multiple thread/CPU contention and shown not to break.

  --

  ZFS bug report: https://github.com/zfsonlinux/zfs/issues/7691

  "I am using LXD containers that are configured to use a ZFS storage backend.
  I create many containers using a benchmark tool, which probably stresses the 
use of ZFS.
  In two out of four attempts, I got

  [  725.970508] INFO: task lxd:4455 blocked for more than 120 seconds.
  [  725.976730]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  725.983551] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  725.991624] INFO: task txg_sync:4202 blocked for more than 120 seconds.
  [  725.998264]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.005071] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.013313] INFO: task lxd:99919 blocked for more than 120 seconds.
  [  726.019609]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.026418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.034560] INFO: task zfs:100513 blocked for more than 120 seconds.
  [  726.040936]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.047746] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.
  [  726.055791] INFO: task zfs:100584 blocked for more than 120 seconds.
  [  726.062170]   Tainted: P   O 4.15.0-20-generic #21-Ubuntu
  [  726.068979] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables 
this message.

  Describe how to reproduce the problem

  Start an Ubuntu 18.04 LTS server.
  Install LXD if not already installed.

  sudo apt update
  sudo apt install lxd lxd-client lxd-tools zfsutils-linux

  Configure LXD with sudo lxd init. When prompted for the storage
  backend, select ZFS and specify an empty disk.

  $ sudo lxd init
  Would you like to use LXD clustering? (yes/no) [default=no]:
   Do you want to configure a new storage pool? (yes/no) [default=yes]:
   Name of the new storage pool [default=default]:
   Name of the storage backend to use (dir, zfs) [default=zfs]:
   Create a new ZFS pool? (yes/no) [default=yes]:
   Would you like to use an existing block device? (yes/no) [default=no]: yes
   Path to the existing block device: /dev/sdb
   Would you like to connect to a MAAS server? (yes/no) [default=no]:
   Would you like to create a new local network bridge? (yes/no) [default=yes]: 
no
   Would you like to configure LXD to use an existing bridge or host interface? 
(yes/no) [default=no]: no
   Would you like LXD to be available over the network? (yes/no) [default=no]:
   Would you like stale cached images to be updated automatically? (yes/no) 
[default=yes]
   Would you like a YAML "lxd init" preseed to be printed? (yes/no) 
[default=no]:

  Now run the following to launch 48 containers in batches of 12.

  lxd-benchmark launch --count 48 --parallel 12

  In two out of four attempts, I got the kernel errors.

  I also tried

  echo 1 >/sys/module/spl/parameters/spl_taskq_kick

  but did not manage to 

[Kernel-packages] [Bug 1769937] Re: Error reported when creating ZFS pool with "-t" option, despite successful pool creation

2018-07-17 Thread Colin Ian King
** Changed in: zfs-linux (Ubuntu)
   Status: Confirmed => In Progress

** Changed in: zfs-linux (Ubuntu)
   Importance: Undecided => High

** Changed in: zfs-linux (Ubuntu)
   Importance: High => Medium

** Changed in: zfs-linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1769937

Title:
  Error reported when creating ZFS pool with "-t" option, despite
  successful pool creation

Status in zfs-linux package in Ubuntu:
  In Progress

Bug description:
  When creating a pool with the -t option on Ubuntu 18.04 LTS, it will
  report an error and return a non-zero exit code, but the pool appears
  to be successfully created.

  For example:

  # zpool create -O canmount=off -O mountpoint=none -R /mnt/rpool-25582 -t 
rpool-25582 rpool /dev/mapper/loop2p1
  cannot create 'rpool': no such pool or dataset
  # zpool list
  NAME  SIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  
ALTROOT
  rpool-25582  69.5G  92.5K  69.5G - 0% 0%  1.00x  ONLINE  
/mnt/rpool-25582
  # zpool export rpool-25582
  # zpool import rpool
  # zpool list
  NAMESIZE  ALLOC   FREE  EXPANDSZ   FRAGCAP  DEDUP  HEALTH  ALTROOT
  rpool  69.5G   134K  69.5G - 0% 0%  1.00x  ONLINE  -

  I opened an upstream bug against ZFS on Linux, here:
  https://github.com/zfsonlinux/zfs/issues/7502

  And it has been recently fixed, here:
  https://github.com/zfsonlinux/zfs/pull/7509

  Can this upstream fix be pulled into the Ubuntu 18.04 LTS packages?

  Thanks,
  Prakash

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1769937/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1742698] Re: zfs-dkms 0.6.5.6-0ubuntu18: zfs kernel module failed to build

2018-07-17 Thread Colin Ian King
The zfs DKMS package is no longer required as the ZFS filesystem modules
are now built into the kernel package, so a quick fix is to remove the
zfs-dkms packagage

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1742698

Title:
  zfs-dkms 0.6.5.6-0ubuntu18: zfs kernel module failed to build

Status in zfs-linux package in Ubuntu:
  Confirmed

Bug description:
  Ubuntu MATE 16.04 standard distro.  It looks like there is a common
  defect across many versions that has not been addressed.  If they are
  trying to fix this, whatever they've been doing is not effective.

  ProblemType: Package
  DistroRelease: Ubuntu 16.04
  Package: zfs-dkms 0.6.5.6-0ubuntu18
  ProcVersionSignature: Ubuntu 4.10.0-42.46~16.04.1-generic 4.10.17
  Uname: Linux 4.10.0-42-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl zcommon znvpair
  ApportVersion: 2.20.1-0ubuntu2.15
  Architecture: amd64
  DKMSBuildLog:
   DKMS make.log for zfs-0.6.5.6 for kernel 4.13.0-26-generic (x86_64)
   Thu Jan 11 06:40:09 MST 2018
   make: *** No targets specified and no makefile found.  Stop.
  DKMSKernelVersion: 4.13.0-26-generic
  Date: Thu Jan 11 06:40:12 2018
  InstallationDate: Installed on 2017-12-16 (25 days ago)
  InstallationMedia: Ubuntu-MATE 16.04.2 LTS "Xenial Xerus" - Release amd64 
(20170215)
  PackageVersion: 0.6.5.6-0ubuntu18
  RelatedPackageVersions:
   dpkg 1.18.4ubuntu1.3
   apt  1.2.24
  SourcePackage: zfs-linux
  Title: zfs-dkms 0.6.5.6-0ubuntu18: zfs kernel module failed to build
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1742698/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1714084] Re: package zfsutils-linux 0.6.5.6-0ubuntu17 failed to install/upgrade: subprocess installed post-installation script returned error exit status 1

2018-07-17 Thread Colin Ian King
*** This bug is a duplicate of bug 1742698 ***
https://bugs.launchpad.net/bugs/1742698

** This bug has been marked a duplicate of bug 1742698
   zfs-dkms 0.6.5.6-0ubuntu18: zfs kernel module failed to build

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1714084

Title:
  package zfsutils-linux 0.6.5.6-0ubuntu17 failed to install/upgrade:
  subprocess installed post-installation script returned error exit
  status 1

Status in zfs-linux package in Ubuntu:
  New

Bug description:
  Attempted reinstall of zfsutils-linux

  ProblemType: Package
  DistroRelease: Ubuntu 16.04
  Package: zfsutils-linux 0.6.5.6-0ubuntu17
  ProcVersionSignature: Ubuntu 4.4.0-92.115-generic 4.4.76
  Uname: Linux 4.4.0-92-generic x86_64
  NonfreeKernelModules: zfs zunicode zcommon znvpair zavl
  ApportVersion: 2.20.1-0ubuntu2.10
  Architecture: amd64
  Date: Wed Aug 30 15:49:18 2017
  ErrorMessage: subprocess installed post-installation script returned error 
exit status 1
  InstallationDate: Installed on 2016-07-24 (402 days ago)
  InstallationMedia: Ubuntu-MATE 16.04.1 LTS "Xenial Xerus" - Release amd64 
(20160719)
  RelatedPackageVersions:
   dpkg 1.18.4ubuntu1.2
   apt  1.2.24
  SourcePackage: zfs-linux
  Title: package zfsutils-linux 0.6.5.6-0ubuntu17 failed to install/upgrade: 
subprocess installed post-installation script returned error exit status 1
  UpgradeStatus: No upgrade log present (probably fresh install)
  modified.conffile..etc.default.zfs: [deleted]

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1714084/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1773584] Re: zfs-dkms 0.7.5-1ubuntu16: zfs kernel module failed to build

2018-07-17 Thread Colin Ian King
*** This bug is a duplicate of bug 1742698 ***
https://bugs.launchpad.net/bugs/1742698

** This bug has been marked a duplicate of bug 1742698
   zfs-dkms 0.6.5.6-0ubuntu18: zfs kernel module failed to build

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1773584

Title:
  zfs-dkms 0.7.5-1ubuntu16: zfs kernel module failed to build

Status in zfs-linux package in Ubuntu:
  Confirmed

Bug description:
  this was reported on the first boot after upgrade to 18.04. I am not
  using zfs currently so my fix is just to remove from the modeule list.

  ProblemType: Package
  DistroRelease: Ubuntu 18.04
  Package: zfs-dkms 0.7.5-1ubuntu16
  ProcVersionSignature: Ubuntu 4.15.0-23.25-generic 4.15.18
  Uname: Linux 4.15.0-23-generic x86_64
  NonfreeKernelModules: zfs zunicode zavl icp zcommon znvpair nvidia
  ApportVersion: 2.20.9-0ubuntu7
  Architecture: amd64
  DKMSBuildLog:
   DKMS make.log for zfs-0.7.5 for kernel 4.15.0-23-generic (x86_64)
   Sat May 26 04:11:20 PDT 2018
   make: *** No targets specified and no makefile found.  Stop.
  DKMSKernelVersion: 4.15.0-23-generic
  Date: Sat May 26 04:11:23 2018
  InstallationDate: Installed on 2018-04-04 (52 days ago)
  InstallationMedia: Xubuntu 17.10 "Artful Aardvark" - Release amd64 (20180105)
  PackageArchitecture: all
  PackageVersion: 0.7.5-1ubuntu16
  Python3Details: /usr/bin/python3.6, Python 3.6.5, python3-minimal, 3.6.5-3
  PythonDetails: /usr/bin/python2.7, Python 2.7.15rc1, python-minimal, 
2.7.15~rc1-1
  RelatedPackageVersions:
   dpkg 1.19.0.5ubuntu2
   apt  1.6.1
  SourcePackage: zfs-linux
  Title: zfs-dkms 0.7.5-1ubuntu16: zfs kernel module failed to build
  UpgradeStatus: Upgraded to bionic on 2018-05-26 (0 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1773584/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1708049] Re: package zfsutils-linux 0.6.5.6-0ubuntu17 failed to install/upgrade: package is in a very bad inconsistent state; you should reinstall it before attempting configur

2018-07-17 Thread Colin Ian King
Can you try:

sudo apt-get install --reinstall zfsutils-linux

and see if that helps.

** Changed in: zfs-linux (Ubuntu)
 Assignee: (unassigned) => Colin Ian King (colin-king)

** Changed in: zfs-linux (Ubuntu)
   Status: New => Triaged

** Changed in: zfs-linux (Ubuntu)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1708049

Title:
  package zfsutils-linux 0.6.5.6-0ubuntu17 failed to install/upgrade:
  package is in a very bad inconsistent state; you should  reinstall it
  before attempting configuration

Status in zfs-linux package in Ubuntu:
  Triaged

Bug description:
  Ubuntu 16.04 . When it trys to update ZFS it fails and then  several hours 
later the computer crashes.Reboot still crashes.  Have to go back several 
updates to stop crashing.
  thanks

  R

  ProblemType: Package
  DistroRelease: Ubuntu 16.04
  Package: zfsutils-linux 0.6.5.6-0ubuntu17
  Uname: Linux 4.9.24-040924-lowlatency x86_64
  ApportVersion: 2.20.1-0ubuntu2.10
  Architecture: amd64
  Date: Wed Aug  2 00:34:53 2017
  DuplicateSignature:
   package:zfsutils-linux:0.6.5.6-0ubuntu17
   Processing triggers for mime-support (3.59ubuntu1) ...
   dpkg: error processing package zfsutils-linux (--configure):
package is in a very bad inconsistent state; you should
  ErrorMessage: package is in a very bad inconsistent state; you should  
reinstall it before attempting configuration
  InstallationDate: Installed on 2017-04-14 (109 days ago)
  InstallationMedia: Ubuntu-Studio 16.04.2 LTS "Xenial Xerus" - Release amd64 
(20170215)
  RelatedPackageVersions:
   dpkg 1.18.4ubuntu1.2
   apt  1.2.20
  SourcePackage: zfs-linux
  Title: package zfsutils-linux 0.6.5.6-0ubuntu17 failed to install/upgrade: 
package is in a very bad inconsistent state; you should  reinstall it before 
attempting configuration
  UpgradeStatus: No upgrade log present (probably fresh install)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1708049/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1778798] Re: zedit doesn't open file in vm mode

2018-07-17 Thread Colin Ian King
zfs-zed is a ZFS Event Daemon and nothing to do with an editor. I think
you must be referring to another tool. Un-assigning this bug from
zfsutils-linux.


** Package changed: zfs-linux (Ubuntu) => fis-gtm (Ubuntu)

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1778798

Title:
  zedit doesn't open file in vm mode

Status in fis-gtm package in Ubuntu:
  New

Bug description:
  fis-gtm installed in Ubuntu environment.  zfs-zed package installed to
  create and edit M routines.  Using zed command doesn't open file.

  --
  Description:Ubuntu 16.04.4 LTS
  Release:16.04
  --
  Package: zfs-zed
  --
  Expected action was the file opened in vm editor.
  --
  Actual action was the file wasn't opened.

  cemberger@cemberger-PC:/$ sudo zed "usr/staff/TESTING"
  cemberger@cemberger-PC:/$ ls usr/staff
  TESTING
  cemberger@cemberger-PC:/$ sudo zed "cprac"
  cemberger@cemberger-PC:/$ sudo zed "ccc"
  cemberger@cemberger-PC:/$

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/fis-gtm/+bug/1778798/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1771091] Re: zpool freezes importing older ZFS pool, blocks shotdown and system does not boot

2018-07-17 Thread Colin Ian King
This looks very similar to the following ZFS bug report

https://github.com/zfsonlinux/zfs/issues/6881

It may indicate you have data corruption and degraded pool. It's hard to
tell any further. Have you been able to resolve this issue?


** Bug watch added: Github Issue Tracker for ZFS #6881
   https://github.com/zfsonlinux/zfs/issues/6881

** Changed in: zfs-linux (Ubuntu)
   Status: Triaged => Incomplete

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1771091

Title:
  zpool freezes importing older ZFS pool, blocks shotdown and system
  does not boot

Status in zfs-linux package in Ubuntu:
  Incomplete

Bug description:
  After fresh install of xubuntu 18.04 LTS 64-bit, 
  and the installation of zfs-dkms I tried to do 'zpool import' on an older ZFS 
pool, consisting of on partition on the separate PATA HDD.

  After issuing 'sudo zpool import ' , command freezes (as to zfs 
commands).
  System then fails to shutdown properly and seems locked and needs hard reboot 
(actually it waits up to half an hour to shutdown).
  After restarting, system displays Xubuntu splash screen and does not boot 
anymore (it actually resets itself if given again half an hour or so).

  When getting to rescue options, by pressing SHIFT key on keyboard and
  going to shell and remounting / read-write, I could do removing of ZFS
  Ubuntu packages and after that system could boot.

  Usefull message I got when trying to continue booting in shell was:
  "[ 40.811792] VERIFY3(0 == remove_reference(hdr,  ((void *)0), tag)) failed 
(0 = 0)  
  [ 40.811856] PANIC at arc.c:3084:arc_buf_destroy()"

  So it points to some ZFS bug with ARC.

  Previously, I was able to (unlike with 17.10) upgrade from 17.10 to 18.04 and 
to import and use a newer ZFS pool.  
  But this bug is about fresh 18.04 install and an older ZFS pool. (zpool 
import says pool can be upgraded)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1771091/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1667838] Re: zpool export mpu1pool cannot export 'mpu1pool': pool is busy

2018-07-17 Thread Colin Ian King
Internally zpool export will call unshare and umount, so a "pool is
busy" indicates that process(es) has/have one or more files open and so
this operation fails.

One can check for these processes using either:

lsof -Pn

or

fuser -vm /poolname

(where poolname is the name of your pool)

and it's worth checking the status of the pool using:

zpool status -v




** Changed in: zfs-linux (Ubuntu)
   Status: In Progress => Incomplete

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1667838

Title:
  zpool export mpu1pool  cannot export 'mpu1pool': pool is busy

Status in zfs-linux package in Ubuntu:
  Incomplete

Bug description:
  $ lsb_release -rd
  Description:  Ubuntu 16.04.2 LTS
  Release:  16.04

  Linux 4.4.0-64-generic

  $ apt-cache policy zfsutils-linux
  zfsutils-linux:
Installed: 0.6.5.6-0ubuntu15

  # modinfo zfs |head
  filename:   /lib/modules/4.4.0-64-generic/kernel/zfs/zfs/zfs.ko
  version:0.6.5.6-0ubuntu15

  
  I was not able to unmount a ZFS pool:

  # zpool export mpu1pool 
  cannot export 'mpu1pool': pool is busy

  The filesystems were unmounted and there was nothing that would be blocking 
them FS.
  Neither sync, echo 3 > /proc/sys/vm/drop_caches helped.

  The only way worked to export a ZFS pool was to turn off the SWAP:

  # lsof -Pn |grep -i mpu1
  lsof: WARNING: can't stat() fuse.gvfsd-fuse file system /run/user/1000/gvfs
Output information may be incomplete.
  # free -mh
totalusedfree  shared  buff/cache   
available
  Mem:   7.7G5.1G502M837M2.1G
1.0G
  Swap:  8.0G428M7.6G

  # swapoff -a 
  # free -mh
totalusedfree  shared  buff/cache   
available
  Mem:   7.7G1.5G4.8G366M1.4G
5.2G
  Swap:0B  0B  0B

  # zpool export mpu1pool 
  # echo $?
  0
  # swapon -a

  
  It is not obvious that the SWAP has to be freed before one can export ZFS 
pool.
  Would it be possible to improve this situation?

  
  Kind regards,
  Andrey Arapov

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1667838/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1723948] Re: error -lock in zfs send

2018-07-17 Thread Colin Ian King
Did the advice in comment #3 help?  I've not heard any response from
this for a couple of months, so I'm closing this bug as Won't Fix. If
this is still and issue, please feel free to reopen the bug.

** Changed in: zfs-linux (Ubuntu)
   Status: Incomplete => Won't Fix

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to zfs-linux in Ubuntu.
https://bugs.launchpad.net/bugs/1723948

Title:
  error -lock in zfs send

Status in Native ZFS for Linux:
  New
Status in zfs-linux package in Ubuntu:
  Won't Fix

Bug description:
  zfs send stop working. the process did not produce output. checked by
  mbuffer log

  in kernel.log: 
  Oct 15 07:25:53 zfs-cis kernel: [479439.151281] INFO: task zfs:8708 blocked 
for more than 120 seconds.
  Oct 15 07:25:53 zfs-cis kernel: [479439.156980]   Tainted: P   OE 
  4.4.0-96-generic #119-Ubuntu
  Oct 15 07:25:53 zfs-cis kernel: [479439.162688] "echo 0 > 
/proc/sys/kernel/hung_task_timeout_secs" disables this message.
  Oct 15 07:25:53 zfs-cis kernel: [479439.173974] zfs D 
88197bd77318 0  8708   8141 0x
  Oct 15 07:25:53 zfs-cis kernel: [479439.173981]  88197bd77318 
810c3dc2 8820374cf000 881b645ff000
  Oct 15 07:25:53 zfs-cis kernel: [479439.173985]  88197bd78000 
00792c6d 882030aa4ac8 
  Oct 15 07:25:53 zfs-cis kernel: [479439.173989]  88101dac1840 
88197bd77330 8183f165 882030aa4a00
  Oct 15 07:25:53 zfs-cis kernel: [479439.173993] Call Trace:
  Oct 15 07:25:53 zfs-cis kernel: [479439.174006]  [] ? 
__wake_up_common+0x52/0x90
  Oct 15 07:25:53 zfs-cis kernel: [479439.174023]  [] 
schedule+0x35/0x80
  Oct 15 07:25:53 zfs-cis kernel: [479439.174045]  [] 
taskq_wait_id+0x60/0xb0 [spl]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174051]  [] ? 
wake_atomic_t_function+0x60/0x60
  Oct 15 07:25:53 zfs-cis kernel: [479439.174115]  [] ? 
dump_write+0x230/0x230 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174178]  [] 
spa_taskq_dispatch_sync+0x92/0xd0 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174223]  [] 
dump_bytes+0x51/0x70 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174267]  [] 
dump_write+0x11e/0x230 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174311]  [] 
backup_cb+0x633/0x850 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174356]  [] 
traverse_visitbp+0x47a/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174365]  [] ? 
spl_kmem_alloc+0xaf/0x190 [spl]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174409]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174462]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174503]  [] 
traverse_dnode+0x7f/0xe0 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174542]  [] 
traverse_visitbp+0x6cc/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174579]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174616]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174653]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174690]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174726]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174762]  [] 
traverse_visitbp+0x5c0/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174799]  [] 
traverse_dnode+0x7f/0xe0 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174835]  [] 
traverse_visitbp+0x865/0x960 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174871]  [] 
traverse_impl+0x1ae/0x410 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174908]  [] ? 
dmu_recv_end_check+0x210/0x210 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174944]  [] 
traverse_dataset+0x52/0x60 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.174981]  [] ? 
dmu_recv_end_check+0x210/0x210 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.175017]  [] 
dmu_send_impl+0x409/0x560 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.175060]  [] 
dmu_send_obj+0x172/0x1e0 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.175129]  [] 
zfs_ioc_send+0xe9/0x2c0 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.175143]  [] ? 
strdup+0x3b/0x60 [spl]
  Oct 15 07:25:53 zfs-cis kernel: [479439.175207]  [] 
zfsdev_ioctl+0x44b/0x4e0 [zfs]
  Oct 15 07:25:53 zfs-cis kernel: [479439.175218]  [] 
do_vfs_ioctl+0x29f/0x490
  Oct 15 07:25:53 zfs-cis kernel: [479439.175225]  [] ? 
_do_fork+0xec/0x360
  Oct 15 07:25:53 zfs-cis kernel: [479439.175232]  [] 
SyS_ioctl+0x79/0x90
  Oct 15 07:25:53 zfs-cis kernel: [479439.175242]  [] 
entry_SYSCALL_64_fastpath+0x16/0x71

To manage notifications about this bug go to:
https://bugs.launchpad.net/zfs/+bug/1723948/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : 

<    7   8   9   10   11   12   13   14   15   16   >