[Kernel-packages] [Bug 1824827] Re: tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
The change introduced by the patch is evident in the kernel message log for Thread 0: between page indexes 2 and 1 there's now another function call to write_cache_pages() instead of just another iteration of the for-loop inside one call. Original kernel: [ 695.276320] Thread 0 running, TID =

[Kernel-packages] [Bug 1824827] Re: tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
Patch posted for SRU: [C][PATCH 0/1] Fix write()/fsync() deadlock in write_cache_pages() https://lists.ubuntu.com/archives/kernel-team/2019-April/100084.html -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu.

[Kernel-packages] [Bug 1824827] Re: tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
testcase test kernel :: latest cosmic version + patch: $ uname -rv 4.18.0-18-generic #19+test20190415b1 SMP Mon Apr 15 15:43:20 UTC 2019 [ 169.145212] kprobe_test: loading out-of-tree module taints kernel. [ 169.149144] kprobe_test: module

[Kernel-packages] [Bug 1824827] Re: tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
testcase, userspace part. ** Attachment added: "test.c" https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1824827/+attachment/5255995/+files/test.c -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu.

[Kernel-packages] [Bug 1824827] Re: tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
testcase original kernel :: latest cosmic version: $ uname -rv 4.18.0-18-generic #19-Ubuntu SMP Tue Apr 2 18:13:16 UTC 2019 [ 654.491029] kprobe_test: loading out-of-tree module taints kernel. [ 654.493322] kprobe_test: module verification

[Kernel-packages] [Bug 1824827] Re: tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
testcase, kernel part. ** Attachment added: "kprobe-test.c" https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1824827/+attachment/5255994/+files/kprobe-test.c -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu.

[Kernel-packages] [Bug 1824827] [NEW] tasks doing write()/fsync() hit deadlock in write_cache_pages()

2019-04-15 Thread Mauricio Faria de Oliveira
fo] * This is only required in Cosmic (for the Bionic HWE kernel), and is already applied in Disco. ** Affects: linux (Ubuntu) Importance: Undecided Status: Invalid ** Affects: linux (Ubuntu Cosmic) Importance: Undecided Assignee: Mauricio Faria de Oliveira (mfo)

[Kernel-packages] [Bug 1821395] Re: fscache: jobs might hang when fscache disk is full

2019-04-12 Thread Mauricio Faria de Oliveira
Regression testing setup/steps === fscache --- sudo apt-get -y install cachefilesd echo 'RUN=yes' | sudo tee -a /etc/default/cachefilesd sudo modprobe fscache sudo systemctl start cachefilesd nfs --- sudo apt-get -y install nfs-kernel-server sudo systemctl start nfs-kernel-server sudo

[Kernel-packages] [Bug 1821395] Re: fscache: jobs might hang when fscache disk is full

2019-04-12 Thread Mauricio Faria de Oliveira
Verification successful with xfstests on nfs+fscache. No regression in bionic-proposed from bionic-updates. bionic-updates / 4.15.0-47: Failures: generic/035 generic/075 generic/091 generic/112 generic/263 generic/294 generic/306 generic/307 generic/430 generic/431 generic/434 generic/469

[Kernel-packages] [Bug 1821395] Re: fscache: jobs might hang when fscache disk is full

2019-04-12 Thread Mauricio Faria de Oliveira
Verification successful with xfstests on nfs+fscache. No regression in cosmic-proposed from cosmic-updates. cosmic-updates / 4.18.0-17: Failures: generic/035 generic/258 generic/294 generic/448 generic/467 generic/477 generic/484 generic/490 generic/495 Failed 9 of 437 tests cosmic-proposed /

[Kernel-packages] [Bug 1821395] Re: fscache: jobs might hang when fscache disk is full

2019-04-11 Thread Mauricio Faria de Oliveira
The verification for bionic/cosmic -proposed is expected to finish by tomorrow (Apr 12). -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1821395 Title: fscache: jobs might hang when

[Kernel-packages] [Bug 1817628] Re: Regular D-state processes impacting LXD containers

2019-03-29 Thread Mauricio Faria de Oliveira
Marking X/B as verification done. The user reports the issue occurs much less often now. Apparently that environment hits some corner case or this may still be expected sometimes under memory pressure (i.e., one big shrinking operation acquired the lock and must finish). Nonetheless, the fix

[Kernel-packages] [Bug 1802021] Re: [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start()

2019-03-25 Thread Mauricio Faria de Oliveira
Updating bug tags to verification done. As mentioned by users in this LP bug, the verification period of 5 days is _usually_ not enough to reproduce this problem, however, we have some datapoints that support the fix is good. 1) The fix has been first delivery in linux-azure, 3 weeks ago, and

[Kernel-packages] [Bug 1821395] Re: fscache: jobs might hang when fscache disk is full

2019-03-22 Thread Mauricio Faria de Oliveira
[B/C][PATCH 0/1] Fix for LP#1821395 (fscache: jobs might hang when fscache disk is full) https://lists.ubuntu.com/archives/kernel-team/2019-March/099448.html ** Description changed: - < NOTE: patches will be sent to kernel-team mailing list. > - [Impact] - * fscache issue where jobs get

[Kernel-packages] [Bug 1821395] Re: fscache: jobs might hang when fscache disk is full

2019-03-22 Thread Mauricio Faria de Oliveira
** Also affects: linux (Ubuntu Cosmic) Importance: Undecided Status: New ** Also affects: linux (Ubuntu Bionic) Importance: Undecided Status: New ** Changed in: linux (Ubuntu) Status: Incomplete => Invalid ** Changed in: linux (Ubuntu Bionic) Status: New =>

[Kernel-packages] [Bug 1821395] [NEW] fscache: jobs might hang when fscache disk is full

2019-03-22 Thread Mauricio Faria de Oliveira
Public bug reported: < NOTE: patches will be sent to kernel-team mailing list. > [Impact] * fscache issue where jobs get hung when fscache disk is full. * trivial upstream fix; already applied in X/D, required in B/C: commit c5a94f434c82 ("fscache: fix race between enablement and

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
[X][PATCH 0/4] LP#1821259 Fix for deadlock in cpu_stopper https://lists.ubuntu.com/archives/kernel-team/2019-March/099427.html [B][PATCH 0/2] Fix for LP#1821259 (pending patches for) Fix for deadlock in cpu_stopper https://lists.ubuntu.com/archives/kernel-team/2019-March/099432.html ** Also

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
Since Bionic already has the fix commit applied, the original kernel version doesn't hit the problem. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1821259 Title: Hard lockup in 2 CPUs

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
Both xenial and bionic original/patched kernels were tested with stress-ng scheduler class, and no regressions were observed. $ stress-ng --version stress-ng, version 0.09.56 (gcc 8.3, x86_64 Linux 4.15.0-47-generic)  $ sudo stress-ng --class scheduler --sequential 0 $ uname -rv

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
Test-case on Xenial; $ ls -1d /sys/devices/system/cpu/cpu[0-9]* /sys/devices/system/cpu/cpu0 /sys/devices/system/cpu/cpu1 Original $ uname -rv 4.4.0-144-generic #170-Ubuntu SMP Thu Mar 14 11:56:20 UTC 2019 $ sudo insmod kmod-stopper/kmod-stopper.ko [ 74.198379] mod_init() :: this

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
-- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1821259 Title: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers Status in linux package in Ubuntu: Incomplete Bug description:

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
Test-case (kmod-stopper.c) - $ sudo apt-get -y install gcc make libelf-dev linux-headers-$(uname -r) $ touch Makefile # fake it, and use this make line: $ make -C /lib/modules/$(uname -r)/build M=$(pwd) obj-m=kmod-stopper.o modules $ echo 9 | sudo tee /proc/sys/kernel/printk $ sudo

[Kernel-packages] [Bug 1821259] Re: Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
Analysis The 1st hard lockup is harder to get the interesting data out of, as apparently the registers with variables related to the cpu number have been clobbered by more recent calls in the spinlock path. Looking at the 2nd hard lockup: addr2line + code shows us that

[Kernel-packages] [Bug 1821259] [NEW] Hard lockup in 2 CPUs due to deadlock in cpu_stoppers

2019-03-21 Thread Mauricio Faria de Oliveira
Public bug reported: [Impact] * This problem hard locks up 2 CPUs in a deadlock, and this soft locks up other CPUs as an effect; the system becomes unusable. * This is relatively rare / difficult to hit because it's a corner case in scheduling/load balancing that needs timing with

[Kernel-packages] [Bug 1817321] Re: installer does not support iSCSI iBFT

2019-03-18 Thread Mauricio Faria de Oliveira
bionic-proposed verification done; iscsi_ibft.ko is present in udeb and loads correctly. --- $ uname -rv 4.15.0-47-generic #50-Ubuntu SMP Wed Mar 13 10:44:52 UTC 2019 $ apt-get download scsi-modules-4.15.0-47-generic-di $ dpkg-deb -c scsi-modules-4.15.0-47-generic-di_4.15.0-47.50_amd64.udeb |

[Kernel-packages] [Bug 1817321] Re: installer does not support iSCSI iBFT

2019-03-18 Thread Mauricio Faria de Oliveira
cosmic-proposed verification done; iscsi_ibft.ko is present in udeb and loads correctly. --- $ uname -rv 4.18.0-17-generic #18-Ubuntu SMP Wed Mar 13 14:34:40 UTC 2019 $ apt-get download scsi-modules-4.18.0-17-generic-di $ dpkg-deb -c scsi-modules-4.18.0-17-generic-di_4.18.0-17.18_amd64.udeb |

[Kernel-packages] [Bug 1802021] Re: [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start()

2019-03-18 Thread Mauricio Faria de Oliveira
Hi Marius @lazamarius1, Per the kernel.ubuntu.com schedule, the version for Bionic/linux -> Xenial/linux-hwe should land soon. You can verify the version/timestamps for each package/release at the bottom of these pages (the linux-hwe version comes a bit after the corresponding linux version)

[Kernel-packages] [Bug 1802021] Re: [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start()

2019-03-18 Thread Mauricio Faria de Oliveira
@lazamarius1, Actually linux-hwe for Bionic with this fix has just been uploaded. See in https://launchpad.net/ubuntu/+source/linux-hwe Changelog linux-hwe (4.15.0-47.50~16.04.1) xenial; urgency=medium ... * [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start() (LP: #1802021) -

[Kernel-packages] [Bug 1817321] Re: installer does not support iSCSI iBFT

2019-03-15 Thread Mauricio Faria de Oliveira
Installer (non-kernel) patches submitted to Debian for feedback. - disk-detect: https://bugs.debian.org/924675 - partman-iscsi: https://bugs.debian.org/924680 ** Bug watch added: Debian Bug tracker #924675 https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=924675 ** Bug watch added: Debian Bug

[Kernel-packages] [Bug 1793901] Re: kernel oops in bcache module

2019-02-27 Thread Mauricio Faria de Oliveira
For documentation purposes, this is the block device setup configured in the system to debug/test/fix/verify the bug. disk partitioning: # fdisk /dev/sdb # create sdb1 and sbd2, 100G each. bcache backing device: LVM volume in partition 1 # pvcreate /dev/sdb1 # vgcreate vg-test /dev/sdb1 #

[Kernel-packages] [Bug 1793901] Re: kernel oops in bcache module

2019-02-27 Thread Mauricio Faria de Oliveira
trusty-proposed: --- verification successful. executed the reproducer 3x on the system used to debug/test/fix this problem. no bug/oops observed. root@petilil:~# uname -rv 3.13.0-166-generic #216-Ubuntu SMP Thu Feb 7 14:07:53 UTC 2019 root@petilil:~# echo writeback >

[Kernel-packages] [Bug 1793901] Re: kernel oops in bcache module

2019-02-27 Thread Mauricio Faria de Oliveira
xenial-proposed: --- verification successful. executed the reproducer 3x on the system used to debug/test/fix this problem. no bug/oops observed. root@petilil:~# uname -rv 4.4.0-143-generic #169-Ubuntu SMP Thu Feb 7 07:56:38 UTC 2019 root@petilil:~# echo writeback >

[Kernel-packages] [Bug 1793901] Re: kernel oops in bcache module

2019-02-27 Thread Mauricio Faria de Oliveira
bionic-proposed: --- verification successful. executed the reproducer 3x on the system used to debug/test/fix this problem. no bug/oops observed. root@petilil:~# uname -rv 4.15.0-46-generic #49-Ubuntu SMP Wed Feb 6 09:33:07 UTC 2019 root@petilil:~# echo writeback >

[Kernel-packages] [Bug 1793901] Re: kernel oops in bcache module

2019-02-27 Thread Mauricio Faria de Oliveira
cosmic-proposed: --- verification successful. executed the reproducer 3x on the system used to debug/test/fix this problem. no bug/oops observed. root@petilil:~# uname -rv 4.18.0-16-generic #17-Ubuntu SMP Fri Feb 8 00:06:57 UTC 2019 root@petilil:~# echo writeback >

[Kernel-packages] [Bug 1817628] Re: Regular D-state processes impacting LXD containers

2019-02-25 Thread Mauricio Faria de Oliveira
[X/B][PATCH] mm: do not stall register_shrinker() https://lists.ubuntu.com/archives/kernel-team/2019-February/098791.html -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1817628 Title:

[Kernel-packages] [Bug 1817628] Re: Regular D-state processes impacting LXD containers

2019-02-25 Thread Mauricio Faria de Oliveira
Also verified this with stress-ng's memory stressor class. The system gets under memory pressure to the point of calling shrink_slab() 10s to 100s thousand times a second (verified with ftrace; function tracer; shrink_slab filter). $ sudo stress-ng --class memory --all -1 & # echo shrink_slab

[Kernel-packages] [Bug 1817628] [NEW] Regular D-state processes impacting LXD containers

2019-02-25 Thread Mauricio Faria de Oliveira
Public bug reported: [Impact] * Systems running under memory pressure may hit stalls in the order of seconds to minutes in systemd-logind and lxd mount operations (e.g., ZFS backend), which get stuck in D state. * The processes stuck in D state have a common stack trace, (cat

[Kernel-packages] [Bug 1817321] Re: installer does not support iSCSI iBFT

2019-02-22 Thread Mauricio Faria de Oliveira
Test for scsi-modules udeb in Disco: --- Check the original/modified udeb contents: $ dpkg-deb -c scsi-modules-4.19.0-13-generic-di_4.19.0-13.14_amd64.udeb | grep iscsi_ibft.ko $ $ dpkg-deb -c scsi-modules-4.19.0-13-generic-di_4.19.0-13.14+test20190221b1_amd64.udeb | grep iscsi_ibft.ko

[Kernel-packages] [Bug 1817321] [NEW] installer does not support iSCSI iBFT

2019-02-22 Thread Mauricio Faria de Oliveira
a de Oliveira (mfo) Status: Confirmed ** Changed in: linux (Ubuntu) Assignee: (unassigned) => Mauricio Faria de Oliveira (mfo) -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/

[Kernel-packages] [Bug 1817321] Re: installer does not support iSCSI iBFT

2019-02-22 Thread Mauricio Faria de Oliveira
linux patch posted to the kernel-team mailing list: [B/C/D/Unstable][PATCH] UBUNTU: d-i: add iscsi_ibft to scsi-modules https://lists.ubuntu.com/archives/kernel-team/2019-February/098745.html ** Changed in: linux (Ubuntu) Status: New => Confirmed -- You received this bug notification

[Kernel-packages] [Bug 1783152] Re: Enable basic support for Solarflare 8000 series NIC

2019-02-20 Thread Mauricio Faria de Oliveira
** Also affects: debian-installer (Ubuntu Precise) Importance: Undecided Status: New ** Also affects: linux (Ubuntu Precise) Importance: Undecided Status: New ** Also affects: linux-lts-xenial (Ubuntu Precise) Importance: Undecided Status: New ** No longer affects:

[Kernel-packages] [Bug 1802021] Re: [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start()

2019-02-14 Thread Mauricio Faria de Oliveira
Hi @lazamarius1, The fix for linux generic should be applied in the next kernel SRU cycle. The current cycle ends on late February [1]. [1] https://kernel.ubuntu.com/ -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux-azure in

[Kernel-packages] [Bug 1811094] Re: iptables connlimit allows more connections than the limit when using multiple CPUs

2019-02-11 Thread Mauricio Faria de Oliveira
Verification successful on trusty-proposed. Updates kernel (goes above 2000 connections) --- root@petilil:~# uname -a Linux petilil 3.13.0-165-generic #215-Ubuntu SMP Wed Jan 16 11:46:47 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux root@petilil:~# iptables -F root@petilil:~# iptables -A INPUT -p tcp

[Kernel-packages] [Bug 1802021] Re: [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start()

2019-02-08 Thread Mauricio Faria de Oliveira
Hi Marcelo (@mhcerri), We have another user who confirmed the 2 patches submitted for linux- azure also fix the problem on linux(-generic). srcu: Prohibit call_srcu() use under raw spinlocks srcu: Lock srcu_data structure in srcu_gp_start() Could they be submitted for linux as well? Thank

[Kernel-packages] [Bug 1802021] Re: [Hyper-V] srcu: Lock srcu_data structure in srcu_gp_start()

2019-02-04 Thread Mauricio Faria de Oliveira
Marcelo @mhcerri, Would you be able to provide a test kernel for bionic/linux-hwe so that @lazamarius1 can provide test results for -generic? I'll be happy to do that as well if you're short on time right now. (I guess the patchset is the same you posted for linux-azure.) Thanks, Mauricio --

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-22 Thread Mauricio Faria de Oliveira
The reporter verified the Bionic kernel successfully as well. -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1810998 Title: CPU hard lockup with rigorous writes to NVMe drive Status in

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-21 Thread Mauricio Faria de Oliveira
Verification done on Bionic (no errors seen on dmesg). Still waiting on verification by the reporter (different hardware), but this verification shows no regression in dmesg. Same steps as described in the previous comment. ** Tags removed: verification-needed-bionic ** Tags added:

[Kernel-packages] [Bug 1810781] Re: mpt3sas - driver using the wrong register to update a queue index in FW

2019-01-18 Thread Mauricio Faria de Oliveira
Verification done on Bionic, with the HWE kernel in Xenial (i.e., 4.15.0-44.47~16.04.1 per the original reporter's environment) The mpt3sas driver is running correctly -- the sosreport shows the previous kernel had mpt3sas fault_state error messages repeatedly within less than 10 minutes, and

[Kernel-packages] [Bug 1810781] Re: mpt3sas - driver using the wrong register to update a queue index in FW

2019-01-17 Thread Mauricio Faria de Oliveira
Verification done on Cosmic for regression on an older adapter model, I/O stress (iozone) finishes successfully, no errors seen in dmesg. Waiting for verification on Bionic by the reporter. root@dixie:~# fdisk /dev/sdb # create one partition root@dixie:~# mkfs.ext4 /dev/sdb1 root@dixie:~# mount

[Kernel-packages] [Bug 1811094] Re: iptables connlimit allows more connections than the limit when using multiple CPUs

2019-01-17 Thread Mauricio Faria de Oliveira
Verification done on Xenial. - server: root@shuckle:~# uname -a Linux shuckle 4.4.0-142-generic #168-Ubuntu SMP Wed Jan 16 21:00:45 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux root@shuckle:~# iptables -F root@shuckle:~# iptables -A INPUT -p tcp -m tcp --syn --dport -m connlimit

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-17 Thread Mauricio Faria de Oliveira
Verification done on Cosmic (no errors seen on dmesg). Waiting for verification on Bionic by the reporter. root@shuckle:~# uname -a Linux shuckle 4.18.0-14-generic #15-Ubuntu SMP Mon Jan 14 09:01:02 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux root@shuckle:~# fdisk /dev/nvme0n1 # create one

[Kernel-packages] [Bug 1811094] Re: iptables connlimit allows more connections than the limit when using multiple CPUs

2019-01-15 Thread Mauricio Faria de Oliveira
Verification done for Bionic. bionic-proposed: --- - server: root@shuckle:~# uname -a Linux shuckle 4.15.0-44-generic #47-Ubuntu SMP Mon Jan 14 11:26:59 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux - client: root@dixie:~# ruby client.rb 10.230.56.116 6000 3 Connecting to

[Kernel-packages] [Bug 1811094] Re: iptables connlimit allows more connections than the limit when using multiple CPUs

2019-01-15 Thread Mauricio Faria de Oliveira
Verification done for Cosmic. cosmic-proposed: --- - server: root@shuckle:~# uname -a Linux shuckle 4.18.0-14-generic #15-Ubuntu SMP Mon Jan 14 09:01:02 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux - client: root@dixie:~# ruby client.rb 10.230.56.116 6000 3 Connecting to

[Kernel-packages] [Bug 1811094] Re: iptables connlimit allows more connections than the limit when using multiple CPUs

2019-01-14 Thread Mauricio Faria de Oliveira
[SRU T][PATCH 0/3] netfilter: nf_conncount: fix for LP#1811094 https://lists.ubuntu.com/archives/kernel-team/2019-January/097878.html [SRU X][PATCH 0/6] netfilter: nf_conncount: fix for LP#1811094 https://lists.ubuntu.com/archives/kernel-team/2019-January/097698.html [SRU B][PATCH 0/5]

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-11 Thread Mauricio Faria de Oliveira
Patch v2 series posted to the kernel-team mailing list: [SRU B][PATCH v2 0/7] blk-wbt: fix for LP#1810998 https://lists.ubuntu.com/archives/kernel-team/2019-January/097831.html [SRU C][PATCH v2 0/6] blk-wbt: fix for LP#1810998

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-10 Thread Mauricio Faria de Oliveira
** Description changed: [Impact] -  * Users may experience cpu hard lockups when performing -    rigorous writes to NVMe drives. + * Users may experience cpu hard lockups when performing +rigorous writes to NVMe drives. -  * The fix addresses an scheduling issue in the original -    

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-09 Thread Mauricio Faria de Oliveira
Patch series posted to kernel-team mailing list: [SRU B][PATCH 00/13] blk-wbt: fix for LP#1810998 https://lists.ubuntu.com/archives/kernel-team/2019-January/097675.html [SRU C][PATCH 0/8] blk-wbt: fix for LP#1810998 https://lists.ubuntu.com/archives/kernel-team/2019-January/097689.html -- You

[Kernel-packages] [Bug 1811094] Re: iptables connlimit allows more connections than the limit when using multiple CPUs

2019-01-09 Thread Mauricio Faria de Oliveira
** Description changed: - The following iptables connlimit rule can be breached - with a multithreaded client and network device driver, - due to a race in the conncount/connlimit code: + [Impact] - # iptables -A INPUT -p tcp -m tcp --syn --dport \ - -m connlimit --connlimit-above 2000

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-09 Thread Mauricio Faria de Oliveira
** Description changed: - [NOTE] - -  * Patches will be sent to the kernel-team mailing list -    once the test kernel has been verified by the reporter. - [Impact]  * Users may experience cpu hard lockups when performing    rigorous writes to NVMe drives.  * The fix addresses an

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-09 Thread Mauricio Faria de Oliveira
** Description changed: [NOTE]  * Patches will be sent to the kernel-team mailing list    once the test kernel has been verified by the reporter. [Impact]  * Users may experience cpu hard lockups when performing    rigorous writes to NVMe drives.  * The fix addresses an

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-09 Thread Mauricio Faria de Oliveira
** Description changed: [NOTE]  * Patches will be sent to the kernel-team mailing list    once the test kernel has been verified by the reporter. [Impact]  * Users may experience cpu hard lockups when performing    rigorous writes to NVMe drives.  * The fix addresses an

[Kernel-packages] [Bug 1811094] [NEW] iptables connlimit allows more connections than the limit when using multiple CPUs

2019-01-09 Thread Mauricio Faria de Oliveira
DROP NOTE: Patches will be sent to the kernel-team mailing list and more details/testing will be provided later today. ** Affects: linux (Ubuntu) Importance: Undecided Assignee: Mauricio Faria de Oliveira (mfo) Status: Confirmed ** Changed in: linux (Ubuntu) Assignee

[Kernel-packages] [Bug 1810998] Re: CPU hard lockup with rigorous writes to NVMe drive

2019-01-08 Thread Mauricio Faria de Oliveira
** Description changed: [NOTE] - * Patches will be sent to the kernel-team mailing list -once the test kernel has been verified by the reporter. +  * Patches will be sent to the kernel-team mailing list +    once the test kernel has been verified by the reporter. [Impact] - *

[Kernel-packages] [Bug 1810998] [NEW] CPU hard lockup with rigorous writes to NVMe drive

2019-01-08 Thread Mauricio Faria de Oliveira
mportance: Undecided Assignee: Mauricio Faria de Oliveira (mfo) Status: Incomplete ** Tags: bionic ** Changed in: linux (Ubuntu) Assignee: (unassigned) => Mauricio Faria de Oliveira (mfo) -- You received this bug notification because you are a member of Kernel Packages, which

[Kernel-packages] [Bug 1810781] Re: mpt3sas - driver using the wrong register to update a queue index in FW

2019-01-07 Thread Mauricio Faria de Oliveira
Patch submitted to kernel-team mailing list, got 2 ACKs. https://lists.ubuntu.com/archives/kernel-team/2019-January/097471.html -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1810781

[Kernel-packages] [Bug 1810781] Re: mpt3sas - driver using the wrong register to update a queue index in FW

2019-01-07 Thread Mauricio Faria de Oliveira
** Description changed: [Impact] * Adapter resets periodically during high-load activity. * I/O stalls until reset/reinit is complete (latency) and I/O performance degrades across cluster (e.g., low throughput from data spread over nodes). * The mpt3sas driver relies in a FW

[Kernel-packages] [Bug 1810781] Re: mpt3sas - driver using the wrong register to update a queue index in FW

2019-01-07 Thread Mauricio Faria de Oliveira
** Description changed: [Impact] + + * Adapter resets periodically during high-load activity. + + * I/O stalls until reset/reinit is complete (latency) and I/O performance + degrades across cluster (e.g., low throughput from data spread over nodes). * The mpt3sas driver relies in a FW

[Kernel-packages] [Bug 1800877] Re: iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-11-14 Thread Mauricio Faria de Oliveira
autopkgtests for open-vm-tools, failures in amd64 and i386. These are consistently failing since May 14th, 2018 for a lot of other packages, thus not a change from this upload. Thanks, Mauricio -- You received this bug notification because you are a member of Kernel Packages, which is

[Kernel-packages] [Bug 1800877] Re: iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-11-14 Thread Mauricio Faria de Oliveira
autopkgtests for ubuntu-fan (triggered by this iproute2 upload) The test failure in ubuntu-fan's lxd test (the thing that is failing in these autopkgtests) does look flaky. Looking at autopkgtest logs for ubuntu-fan on xenial, there's a series of retriggers for the _same_ version of docker.io

[Kernel-packages] [Bug 1797990] Re: kdump fail due to an IRQ storm

2018-11-08 Thread Mauricio Faria de Oliveira
Attaching for documentation purposes, Tarball with 'dmesg -t | sort' for boot/kexec & option disabled/enabled, in Xenial, Bionic, Cosmic, Disco. ** Attachment added: "sf202166.dmesg.tar.xz"

[Kernel-packages] [Bug 1797990] Re: kdump fail due to an IRQ storm

2018-11-08 Thread Mauricio Faria de Oliveira
Patch set v2 submitted to the kernel-team mailing list for Xenial, Bionic, Cosmic, Disco. [SRU X][PATCH v2 0/3] Add kernel parameter 'pci=clearmsi' to clear MSI(X)s early on boot https://lists.ubuntu.com/archives/kernel-team/2018-November/096631.html [SRU B][PATCH v2 0/3] Add kernel parameter

[Kernel-packages] [Bug 1797990] Re: kdump fail due to an IRQ storm

2018-11-08 Thread Mauricio Faria de Oliveira
** Description changed: [Impact]  * A kexec/crash kernel might get stuck and fail to boot    (for crash kernel, kdump fails to collect a crashdump)    if a PCI device is buggy/stuck/looping and triggers a    continuous flood of MSI(X) interrupts (that the kernel    does not yet

[Kernel-packages] [Bug 1797990] Re: kdump fail due to an IRQ storm

2018-11-07 Thread Mauricio Faria de Oliveira
** Description changed: [Impact]  * A kexec/crash kernel might get stuck and fail to boot    (for crash kernel, kdump fails to collect a crashdump)    if a PCI device is buggy/stuck/looping and triggers a    continuous flood of MSI(X) interrupts (that the kernel    does not yet

[Kernel-packages] [Bug 1797990] Re: kdump fail due to an IRQ storm

2018-11-07 Thread Mauricio Faria de Oliveira
** Description changed: + [Impact] + + * A kexec/crash kernel might get stuck and fail to boot +(for crash kernel, kdump fails to collect a crashdump) +if a PCI device is buggy/stuck/looping and triggers a +continuous flood of MSI(X) interrupts (that the kernel +does not yet

[Kernel-packages] [Bug 1800877] Re: iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-11-07 Thread Mauricio Faria de Oliveira
Verification successful on xenial-proposed. Updated verification tags. Steps = Setup 1) Enable xenial-proposed and install the iproute2 package: --- $ echo 'deb http://archive.ubuntu.com/ubuntu xenial-proposed main restricted' | sudo tee /etc/apt/sources.list.d/xenial-proposed.list deb

[Kernel-packages] [Bug 1800877] Re: iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-11-02 Thread Mauricio Faria de Oliveira
Hi Eric, This is the v3 debdiff with the patches updated to apply with no offset messages per your request. Thanks, Mauricio -- Applying patch debian/patches/1008-vf_trust_dddf1b44126e.patch patching file include/linux/if_link.h Applying patch debian/patches/1009-vf_trust_b6d77d9ee312.patch

[Kernel-packages] [Bug 1800877] Re: iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-11-01 Thread Mauricio Faria de Oliveira
Hi Eric, Thanks for reviewing. This is the debdiff v2, addressing the points you brought up (numbered patches, bug-ubuntu dep3 tag). cheers, Mauricio ** Patch removed: "iproute2_xenial_vf-trust.debdiff"

[Kernel-packages] [Bug 1800877] Re: iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-10-31 Thread Mauricio Faria de Oliveira
** Patch added: "iproute2_xenial_vf-trust.debdiff" https://bugs.launchpad.net/ubuntu/+source/iproute2/+bug/1800877/+attachment/5207579/+files/iproute2_xenial_vf-trust.debdiff -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to iproute2 in

[Kernel-packages] [Bug 1800877] [NEW] iproute2/xenial: Add support for the VF Trust setting (fix IPv6 multicast under SR-IOV on Mellanox adapters)

2018-10-31 Thread Mauricio Faria de Oliveira
Public bug reported: [Impact] * An VM's VF cannot receive IPv6 multicast traffic from other VMs' VFs in the same Mellanox adapter _if_ its VF trust setting is not enabled, and on Xenial currently iproute2 _cannot_ enable it. * This breaks IPv6 NDP (Neighbor Discovery Protocol) in

[Kernel-packages] [Bug 1798110] Re: xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-25 Thread Mauricio Faria de Oliveira
Verification successful with the kernel in xenial-proposed. Updating verification tags. Details === Guest: --- root@mfo-sf194614:~# apt-cache madison linux-image-4.4.0-139-generic linux-image-4.4.0-139-generic | 4.4.0-139.165 | http://archive.ubuntu.com/ubuntu/ xenial-proposed/main amd64

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12

2018-10-24 Thread Mauricio Faria de Oliveira
Verification successful on Cosmic. --- $ apt-cache madison linux-image-4.18.0-11-generic linux-image-4.18.0-11-generic | 4.18.0-11.12 | http://archive.ubuntu.com/ubuntu cosmic-proposed/main amd64 Packages $ sudo apt-get install -y linux-image-4.18.0-11-generic=4.18.0-11.12 $ sudo reboot $

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12

2018-10-24 Thread Mauricio Faria de Oliveira
Verification successful on Bionic. --- $ apt-cache madison linux-image-4.15.0-39-generic linux-image-4.15.0-39-generic | 4.15.0-39.42 | http://archive.ubuntu.com/ubuntu bionic-proposed/main amd64 Packages $ sudo apt-get install -y linux-image-4.15.0-39-generic=4.15.0-39.42 $ sudo reboot $

[Kernel-packages] [Bug 1798110] Re: xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-16 Thread Mauricio Faria de Oliveira
Synthetic Test Case === Create a Xenial KVM guest --- mfo@rotom:~$ uvt-simplestreams-libvirt sync release=xenial arch=amd64 mfo@rotom:~$ uvt-kvm create --mem 4096 --cpu 2 --disk 16 mfo-sf194614 release=xenial arch=amd64 Modify the virsh XML to add a virtio-scsi controller +

[Kernel-packages] [Bug 1798110] Re: xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-16 Thread Mauricio Faria de Oliveira
Organic Testcase set -x instancename=USER-trusty firstdevice=sdb seconddevice=sdc diskname=USER-disk-1 gcloud compute --project "PROJECT" ssh --zone "ZONE" "$instancename" --command '(tar -C /var/issue/first -cf - .) | (tar -C /var/issue/second -xpf -)' gcloud compute --project

[Kernel-packages] [Bug 1798110] Re: xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-16 Thread Mauricio Faria de Oliveira
Patches posted to kernel-team mailing list [1]. [1] https://lists.ubuntu.com/archives/kernel-team/2018-October/096072.html [SRU Xenial][PATCH 0/2] Improve our SAUCE for virtio-scsi reqs counter (fix CPU soft lockup) ** Description changed: [Impact] - * Detaching virtio-scsi disk in

[Kernel-packages] [Bug 1798110] Re: xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-16 Thread Mauricio Faria de Oliveira
Organic Testcase: It usually reproduces the problem in the 1st or 2nd iteration. With the fix the problem did not reproduce in 35 iterations. ** Description changed: - (I'll add the SRU template + testing steps and post to ML shortly.) + [Impact] + + * Detaching virtio-scsi disk in Xenial

[Kernel-packages] [Bug 1798110] Re: xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-16 Thread Mauricio Faria de Oliveira
Problem Analysis The dmesg log 'crash/201809061748/dmesg.201809061748' shows the CPU soft lockup occurs 25 seconds after the 'sdb' virtio-scsi drive is removed. This seems to indicate the events are related (there's usually an extra 2s-3s between an event and the report of the

[Kernel-packages] [Bug 1798110] [NEW] xenial: virtio-scsi: CPU soft lockup due to loop in virtscsi_target_destroy()

2018-10-16 Thread Mauricio Faria de Oliveira
Public bug reported: (I'll add the SRU template + testing steps and post to ML shortly.) A customer reported a CPU soft lockup on Trusty HWE kernel from Xenial when detaching a virtio-scsi drive, and provided a crashdump that shows 2 things: 1) The soft locked up CPU is waiting for another CPU

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12

2018-10-08 Thread Mauricio Faria de Oliveira
test on Bionic -- original: $ uname -a Linux mfo-sf194734 4.15.0-37-generic #40-Ubuntu SMP Tue Oct 2 15:36:54 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux $ bridge fdb show 33:33:00:00:00:01 dev ens3 self permanent 01:00:5e:00:00:01 dev ens3 self permanent

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12

2018-10-08 Thread Mauricio Faria de Oliveira
** Changed in: linux (Ubuntu) Status: Incomplete => Confirmed -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1796748 Title: regression in 'ip --family bridge neigh' since linux

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12

2018-10-08 Thread Mauricio Faria de Oliveira
** Description changed: [Impact] - * Netlink RTM_GETNEIGH requests for PF_BRIDGE are broken since linux +  * Netlink RTM_GETNEIGH requests for PF_BRIDGE are broken since linux v4.12. - * Users, tools (e.g., iproute2), and libraries (e.g., go netlink) that use -such request/family

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12

2018-10-08 Thread Mauricio Faria de Oliveira
** Summary changed: - regression in 'ip --family bridge neigh' since linux v4.12+ + regression in 'ip --family bridge neigh' since linux v4.12 -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu.

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12+

2018-10-08 Thread Mauricio Faria de Oliveira
test on unstable: - original: $ uname -a Linux mfo-sf194734 4.19.0-1-generic #2 SMP Mon Oct 8 14:38:01 -03 2018 x86_64 x86_64 x86_64 GNU/Linux $ bridge fdb show 33:33:00:00:00:01 dev ens3 self permanent 01:00:5e:00:00:01 dev ens3 self permanent

[Kernel-packages] [Bug 1796748] Re: regression in 'ip --family bridge neigh' since linux v4.12+

2018-10-08 Thread Mauricio Faria de Oliveira
test on Cosmic -- original: $ uname -a Linux mfo-sf194734 4.18.0-9-generic #10-Ubuntu SMP Fri Oct 5 16:49:44 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux $ bridge fdb show 33:33:00:00:00:01 dev ens3 self permanent 01:00:5e:00:00:01 dev ens3 self permanent

[Kernel-packages] [Bug 1796748] [NEW] regression in 'ip --family bridge neigh' since linux v4.12

2018-10-08 Thread Mauricio Faria de Oliveira
Public bug reported: [Impact] * Netlink RTM_GETNEIGH requests for PF_BRIDGE are broken since linux v4.12. * Users, tools (e.g., iproute2), and libraries (e.g., go netlink) that use such request/family currently receive nothing back in the kernel response. * The upstream fix resolves the

[Kernel-packages] [Bug 1787281] Re: errors when scanning partition table of corrupted AIX disk

2018-09-17 Thread Mauricio Faria de Oliveira
SRU Verification: all 3 stable releases successfully survive the disk probe/partition scan of the disk image with corrupted AIX partition table. Updating verification tags. Thanks! Trusty: --- $ uname -a Linux trusty 3.13.0-159-generic #209-Ubuntu SMP Wed Sep 12 10:01:59 UTC 2018 x86_64

[Kernel-packages] [Bug 1783152] Re: Enable basic support for Solarflare 8000 series NIC

2018-09-10 Thread Mauricio Faria de Oliveira
The debian-installer images in trusty-proposed work correctly; changing verification tags to done. I tested in the following platforms, with plain and LVM partitioning (details in comment #10). - amd64 bare-metal & qemu-kvm guest - i386 qemu-kvm guest - arm64 qemu guest - ppc64el qemu guest The

[Kernel-packages] [Bug 1783152] Re: Enable basic support for Solarflare 8000 series NIC

2018-09-06 Thread Mauricio Faria de Oliveira
The secure boot packages on trusty-proposed are now correct (LP: #1708245), and the d-i rebuild on amd64 is in progress (verified successfully on PPA). -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu.

[Kernel-packages] [Bug 1783152] Re: Enable basic support for Solarflare 8000 series NIC

2018-09-04 Thread Mauricio Faria de Oliveira
@slashd thanks for the pointers. Yesterday/today I worked on understanding and providing a workaround for this, but a more complete solution (more updates to secure-boot related packages) is already in the works (in cyphermox's PPAs) and should be rolled out soon, according to him. Once that

[Kernel-packages] [Bug 1783152] Re: Enable basic support for Solarflare 8000 series NIC

2018-08-30 Thread Mauricio Faria de Oliveira
This is the patch for trusty debian-installer to pick up this new xenial hwe kernel. It's been tested by the customer with the SF 8000 series NICs on amd64 bare metal (back when using test packages from their private PPA), so the netboot install works on that NIC model. I built it on PPA for the

  1   2   >