[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
This bug was fixed in the package linux - 3.13.0-158.208 --- linux (3.13.0-158.208) trusty; urgency=medium * linux: 3.13.0-158.208 -proposed tracker (LP: #1788764) * CVE-2018-3620 // CVE-2018-3646 - SAUCE: x86/fremap: Invert the offset when converting to/from a PTE * BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) (LP: #1780470) - VSOCK: sock_put wasn't safe to call in interrupt context - VSOCK: Fix lockdep issue. - VSOCK: Detach QP check should filter out non matching QPs. * CacheFiles: Error: Overlong wait for old active object to go away. (LP: #1776254) - cachefiles: Fix missing clear of the CACHEFILES_OBJECT_ACTIVE flag - cachefiles: Wait rather than BUG'ing on "Unexpected object collision" * fscache cookie refcount updated incorrectly during fscache object allocation (LP: #1776277) - fscache: Fix reference overput in fscache_attach_object() error handling * FS-Cache: Assertion failed: FS-Cache: 6 == 5 is false (LP: #1774336) - Revert "UBUNTU: SAUCE: CacheFiles: fix a read_waiter/read_copier race" - fscache: Allow cancelled operations to be enqueued - cachefiles: Fix refcounting bug in backing-file read monitoring -- Kleber Sacilotto de Souza Fri, 24 Aug 2018 15:08:23 + ** Changed in: linux (Ubuntu Trusty) Status: Fix Committed => Fix Released ** CVE added: https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-3620 ** CVE added: https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-3646 -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: Fix Released Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173]
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
Here's what has been brought to my attention by an impacted user : "No news is good news, no crashes, no abnormal behaviour ... I'll let you know if this changes but I suspect we're good to ship it! Thanks for all your help. " Please also note that a test kernel ran for this same impacted user for about 4 weeks without crashes or unwanted situation on VMware < & > 6.5. - Eric ** Tags removed: verification-needed-trusty ** Tags added: verification-done-trusty -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: Fix Committed Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Test Case] - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. [Regression Potential] * A test kernel has been tested on VMware 5.5 (where the situation doesn't happen but we are testing anyway in order to look for potential regression in earlier VMware environment version) and Vmware 6.5 (where the situation happens). I basically wanted to make sure the test package will work good on both VMware env before thinking of proposing the patch series to the Ubuntu kernel ML. And it does so far, see Comment #5 from an affected
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
The trusty-proposed package ^ : # apt-get changelog linux-image-3.13.0-158-generic ... * BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) (LP: #1780470) - VSOCK: sock_put wasn't safe to call in interrupt context - VSOCK: Fix lockdep issue. - VSOCK: Detach QP check should filter out non matching QPs. ... # apt-cache policy linux-image-3.13.0-158-generic linux-image-3.13.0-158-generic: Installed: (none) Candidate: 3.13.0-158.208 Version table: 3.13.0-158.208 0 500 http://archive.ubuntu.com/ubuntu/ trusty-proposed/main amd64 Packages -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: Fix Committed Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Test Case] - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. [Regression Potential] * A test kernel has been tested on VMware 5.5 (where the situation doesn't happen but we are testing anyway in order to look for potential regression in earlier VMware environment version) and Vmware 6.5 (where the situation happens). I basically wanted to make sure the test package will work good on both VMware
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
This bug is awaiting verification that the kernel in -proposed solves the problem. Please test the kernel and update this bug with the results. If the problem is solved, change the tag 'verification-needed- trusty' to 'verification-done-trusty'. If the problem still exists, change the tag 'verification-needed-trusty' to 'verification-failed- trusty'. If verification is not done by 5 working days from today, this fix will be dropped from the source code, and this bug will be closed. See https://wiki.ubuntu.com/Testing/EnableProposed for documentation how to enable and use -proposed. Thank you! ** Tags added: verification-needed-trusty -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: Fix Committed Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Test Case] - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. [Regression Potential] * A test kernel has been tested on VMware 5.5 (where the situation doesn't happen but we are testing anyway in order to look for potential regression in earlier VMware environment version) and Vmware 6.5 (where the situation happens). I basically wanted to make sure the
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
** Changed in: linux (Ubuntu Trusty) Status: In Progress => Fix Committed -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: Fix Committed Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Test Case] - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. [Regression Potential] * A test kernel has been tested on VMware 5.5 (where the situation doesn't happen but we are testing anyway in order to look for potential regression in earlier VMware environment version) and Vmware 6.5 (where the situation happens). I basically wanted to make sure the test package will work good on both VMware env before thinking of proposing the patch series to the Ubuntu kernel ML. And it does so far, see Comment #5 from an affected user. * Limited risk due to changes being made only to a very specific driver. [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Test Case] - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. [Regression Potential] - A test kernel has been tested on VMware 5.5 (where the situation doesn't - happen but we are testing anyway in order to look for potential + * A test kernel has been tested on VMware 5.5 (where the situation + doesn't happen but we are testing anyway in order to look for potential regression in earlier VMware environment version) and Vmware 6.5 (where the situation happens). I basically wanted to make sure the test package will work good on both VMware env before thinking of proposing the patch series to the Ubuntu kernel ML. And it does so far, see Comment #5 from an affected user. + + * Limited risk due to changes being made only to a + very specific driver. + [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch series 8ab18d7 VSOCK: Detach QP check should filter out non matching QPs. 8566b86 VSOCK: Fix lockdep issue. 4ef7ea9 VSOCK: sock_put wasn't safe to call in interrupt context commit 4ef7ea9 Author: Jorgen Hansen Date: Wed Oct 21
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
Feedback from an affected user (3 weeks after he start installing a test kernel with the potential candidate fixes : " ..we are running the test kernel on 13 VMs, 10 on the new environment (VMware 6.5) and 3 on the old (VMware 5.5). We have experienced no hangs for those VMs in the new environment, have been installing the test kernels on more VMs as they hang, and have experienced no ill effects in the old environment... " With that being said, I'll submit the patch series to the Ubuntu kernel ML for final approval, when the next SRU cycle will be announce. ** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Test Case] - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. [Regression Potential] - A test kernel is now being tested on VMware 5.5 (where the situation - doesn't happen but we are testing anyway in order to look for potential + A test kernel has been tested on VMware 5.5 (where the situation doesn't + happen but we are testing anyway in order to look for potential regression in earlier VMware environment version) and Vmware 6.5 (where - the situation happens). We basically want to make sure the test package + the situation happens). I basically wanted to make sure the test package will work good on both VMware env before thinking of proposing the patch series to the Ubuntu kernel ML. + + And it does so far, see Comment #5 from an affected user. [Other
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! + [Test Case] + - Run a Trusty kernel (3.13 kernel series) VMware guests hosted on VMware 6.0 and later (with VMCI (Virtual Machine Communication Interface) turn on) + - Wait for the crash to happen. It is happening randomly with no clear reproducible steps on how to trigger the crash for now. + + [Regression Potential] + + A test kernel is now being tested on VMware 5.5 (where the situation + doesn't happen but we are testing anyway in order to look for potential + regression in earlier VMware environment version) and Vmware 6.5 (where + the situation happens). We basically want to make sure the test package + will work good on both VMware env before thinking of proposing the patch + series to the Ubuntu kernel ML. + [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch series 8ab18d7 VSOCK: Detach QP check should filter out non matching QPs. 8566b86 VSOCK: Fix lockdep issue. 4ef7ea9 VSOCK: sock_put wasn't safe to call in interrupt context commit 4ef7ea9 Author: Jorgen Hansen Date: Wed Oct 21 04:53:56 2015 -0700 VSOCK: sock_put wasn't safe to call in interrupt context ... Multiple customers have been hitting this issue when using VMware tools on vSphere 2015. ... VSphere 2015 == VMware 6.0 (release in 2015) and late.
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
I currently have a user who run the test kernel (ppa:slashd/sf185584) on both of his environment (Vmware 5.5 and VMware 6.5), to see if it fixes the situation on 6.5 with v3.13 guests and to make sure everything still looks good on VMware 5.5 (to avoid potential regression on version < 6.5) -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: In Progress Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch series 8ab18d7 VSOCK: Detach QP check should filter out non matching QPs. 8566b86 VSOCK: Fix lockdep issue. 4ef7ea9 VSOCK: sock_put wasn't safe to call in interrupt context commit 4ef7ea9 Author: Jorgen Hansen Date: Wed Oct 21 04:53:56 2015 -0700 VSOCK: sock_put wasn't safe to call in interrupt context ... Multiple customers have been hitting this issue when using VMware tools on vSphere 2015. ... VSphere 2015 == VMware 6.0 (release in 2015) and late. [3] - https://patchwork.kernel.org/patch/9948741/ To manage notifications about this bug go to:
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
This user will run the test kernel for several weeks to have a good sample. ** Tags added: sts ** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. - - The crashes started after the above move (5.5->6.5). Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch series 8ab18d7 VSOCK: Detach QP check should filter out non matching QPs. 8566b86 VSOCK: Fix lockdep issue. 4ef7ea9 VSOCK: sock_put wasn't safe to call in interrupt context commit 4ef7ea9 Author: Jorgen Hansen Date: Wed Oct 21 04:53:56 2015 -0700 VSOCK: sock_put wasn't safe to call in interrupt context ... Multiple customers have been hitting this issue when using VMware tools on vSphere 2015. ... VSphere 2015 == VMware 6.0 (release in 2015) and late. [3] - https://patchwork.kernel.org/patch/9948741/ ** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
I built a test kernel[1] (3.13.0-153.203+hf185584v20180706b4)on my PPA containing the 3 fixes as suggested by the VMware author[2]. If you are affected by this problem, feel free to use it for testing and update this bug with any feedbacks. [1] - PPA sudo add-apt-repository ppa:slashd/sf185584 sudo apt-get update [2] - https://patchwork.kernel.org/patch/9948741/ -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to linux in Ubuntu. https://bugs.launchpad.net/bugs/1780470 Title: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late) Status in linux package in Ubuntu: Fix Released Status in linux source package in Trusty: In Progress Bug description: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. The crashes started after the above move (5.5->6.5). Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch series 8ab18d7 VSOCK: Detach QP check should filter out non matching QPs. 8566b86 VSOCK: Fix lockdep issue. 4ef7ea9 VSOCK: sock_put wasn't safe to call in interrupt context commit 4ef7ea9 Author: Jorgen Hansen Date: Wed Oct 21 04:53:56 2015 -0700 VSOCK: sock_put wasn't safe to call in interrupt context ... Multiple customers have been hitting this issue when using VMware tools on vSphere 2015. ... VSphere 2015 == VMware 6.0 (release
[Kernel-packages] [Bug 1780470] Re: BUG: scheduling while atomic (Kernel : Ubuntu-3.13 + VMware: 6.0 and late)
** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. The crashes started after the above move (5.5->6.5). Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while atomic: swapper/3/0/0x0100 [17007961.189794] Modules linked in: arc4 md4 nls_utf8 cifs nfsv3 nfs_acl nfsv4 nfs lockd sunrpc fscache veth ipt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 xt_addrtype iptable_filter ip_tables xt_conntrack x_tables nf_nat nf_conntrack bridge stp llc aufs vmw_vsock_vmci_transport vsock ppdev vmwgfx serio_raw coretemp ttm drm vmw_balloon vmw_vmci shpchp i2c_piix4 parport_pc mac_hid xfs lp libcrc32c parport psmouse floppy vmw_pvscsi vmxnet3 pata_acpi [17007961.189856] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 3.13.0-135-generic #184-Ubuntu [17007961.189862] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 10/22/2013 [17007961.189867] 88042f263b90 8172d959 88042f263d30 [17007961.189874] 88042f273180 88042f263ba0 81726d8c 88042f263c00 [17007961.189879] 81731c8f 880428c29800 00013180 880428c25fd8 [17007961.189885] Call Trace: [17007961.189889] [] dump_stack+0x64/0x82 [17007961.189913] [] __schedule_bug+0x4c/0x5a [17007961.189922] [] __schedule+0x6af/0x7f0 [17007961.189929] [] schedule+0x29/0x70 [17007961.189935] [] schedule_timeout+0x279/0x310 [17007961.189947] [] ? select_task_rq_fair+0x56b/0x6f0 [17007961.189955] [] ? enqueue_task_fair+0x422/0x6d0 [17007961.189962] [] ? sched_clock_cpu+0xb5/0x100 [17007961.189971] [] wait_for_completion+0xa6/0x150 [17007961.189977] [] ? wake_up_state+0x20/0x20 [17007961.189987] [] ? __call_rcu+0x2d0/0x2d0 [17007961.189993] [] wait_rcu_gp+0x4b/0x60 [17007961.18] [] ? ftrace_raw_output_rcu_utilization+0x50/0x50 [17007961.190006] [] synchronize_sched+0x3a/0x50 [17007961.190047] [] vmci_event_unsubscribe+0x76/0xb0 [vmw_vmci] [17007961.190063] [] vmci_transport_destruct+0x21/0xe0 [vmw_vsock_vmci_transport] [17007961.190078] [] vsock_sk_destruct+0x17/0x60 [vsock] [17007961.190087] [] __sk_free+0x1f/0x180 [17007961.190092] [] sk_free+0x19/0x20 [17007961.190102] [] vmci_transport_recv_stream_cb+0x200/0x2f0 [vmw_vsock_vmci_transport] [17007961.190114] [] vmci_datagram_invoke_guest_handler+0xbc/0xf0 [vmw_vmci] [17007961.190126] [] vmci_dispatch_dgs+0xcf/0x230 [vmw_vmci] [17007961.190138] [] tasklet_action+0x11e/0x130 [17007961.190145] [] __do_softirq+0xfc/0x310 [17007961.190153] [] irq_exit+0x105/0x110 [17007961.190161] [] do_IRQ+0x56/0xc0 [17007961.190170] [] common_interrupt+0x6d/0x6d [17007961.190173] [] ? native_safe_halt+0x6/0x10 [17007961.190190] [] default_idle+0x1f/0x100 [17007961.190197] [] arch_cpu_idle+0x26/0x30 [17007961.190205] [] cpu_startup_entry+0xc1/0x2b0 [17007961.190214] [] start_secondary+0x21d/0x2d0 [17007961.190221] bad: scheduling from the idle thread! [Other infos] I have identified a patch series[2] which seems to fix the exact same situation. A full discussion can be found on patchworks[3], suggesting a certain patch series[2] authored by Vmware. [1] - VM details : Release: Trusty Kernel: Ubuntu-3.13.0-135 [2] Upstream patch series 8ab18d7 VSOCK: Detach QP check should filter out non matching QPs. 8566b86 VSOCK: Fix lockdep issue. 4ef7ea9 VSOCK: sock_put wasn't safe to call in interrupt context + + + commit 586fbeba3fa97614d9124f8d56b4000c81368ff4 + Author: Jorgen Hansen + Date: Wed Oct 21 04:53:56 2015 -0700 + + VSOCK: sock_put wasn't safe to call in interrupt context + + ... + Multiple customers have been hitting this issue when using + VMware tools on vSphere 2015. + ... + + + VSphere 2015 == VMware 6.0 (release in 2015) and late. + + [3] - https://patchwork.kernel.org/patch/9948741/ ** Description changed: [Impact] It has been brought to my attention that VMware Guest[1] randomly crashes after moving the VMs from a VMware "5.5" env to VMware 6.5 env. The crashes started after the above move (5.5->6.5). Notes: * The crashes wasn't present in VMware 5.5 (with the same VMs). Only started to happens with Vmware 6.5 * The Trusty HWE kernel (Ubuntu-4.4.0-X) doesn't exhibit the situation on VMware 6.5. Here's the stack trace took from the .vmss converted to be readable by Linux debugger: [17007961.187411] BUG: scheduling while