On 9/14/10 5:21 PM, Temple Jason wrote:

Hello,

I have recently upgraded my lustre filesystem from 1.8.3 to 1.8.4. The first day we brought our system online with the new version, we started seeing clients getting stuck in this soft lockup loop. The load shoots up over 120, and eventually the node becomes unusable and requires a hard reset. I've seen loops like this on the server side in previous lustre versions, but to have it happen on the client is completely new. Here is a bit of what I see in the logs:

Sep 13 21:11:39 wn122 kernel: LustreError: 27016:0:(statahead.c:289:ll_sai_entry_fini()) ASSERTION(sa_is_stopped(sai)) failed

The soft lockup is caused by this ASSERTION, because when such ASSERTION triggered, the thread holds a spin_lock, then caused other threads busy waiting for such spin_lock.

Sep 13 21:11:39 wn122 kernel: LustreError: 27016:0:(statahead.c:289:ll_sai_entry_fini()) LBUG

Sep 13 21:11:39 wn122 kernel: Pid: 27016, comm: athena.py

Sep 13 21:11:39 wn122 kernel:

Sep 13 21:11:39 wn122 kernel: Call Trace:

Sep 13 21:11:39 wn122 kernel: [<ffffffff885cf6a1>] libcfs_debug_dumpstack+0x51/0x60 [libcfs]

Sep 13 21:11:39 wn122 kernel: [<ffffffff885cfbda>] lbug_with_loc+0x7a/0xd0 [libcfs]

Sep 13 21:11:39 wn122 kernel: [<ffffffff885d7f00>] tracefile_init+0x0/0x110 [libcfs]

Sep 13 21:11:39 wn122 kernel: [<ffffffff889247e9>] ll_statahead_exit+0x409/0x500 [lustre]

Sep 13 21:11:39 wn122 kernel: [<ffffffff8008a4b4>] default_wake_function+0x0/0xe

Sep 13 21:11:39 wn122 kernel: [<ffffffff888d2f2e>] ll_intent_drop_lock+0x8e/0xb0 [lustre]

Sep 13 21:11:39 wn122 kernel: [<ffffffff8891c2bb>] ll_lookup_it+0x30b/0x7c0 [lustre]

Sep 13 21:11:39 wn122 kernel: [<ffffffff888e250d>] __ll_inode_revalidate_it+0x5bd/0x650 [lustre]

Sep 13 21:11:39 wn122 kernel: [<ffffffff886f34f4>] ldlm_lock_add_to_lru+0x74/0xe0 [ptlrpc]

Sep 13 21:11:39 wn122 kernel: [<ffffffff88919431>] ll_convert_intent+0xb1/0x170 [lustre]

Sep 13 21:11:39 wn122 kernel: [<ffffffff8891d4c7>] ll_lookup_nd+0x207/0x400 [lustre]

Sep 13 21:11:39 wn122 kernel:  [<ffffffff8002229b>] d_alloc+0x174/0x1a9

Sep 13 21:11:39 wn122 kernel:  [<ffffffff8000cc40>] do_lookup+0xe5/0x1e6

Sep 13 21:11:39 wn122 kernel: [<ffffffff80009fce>] __link_path_walk+0xa01/0xf42

Sep 13 21:11:39 wn122 kernel: [<ffffffff8000e803>] link_path_walk+0x5c/0xe5

Sep 13 21:11:39 wn122 kernel:  [<ffffffff80034cd4>] vfs_readdir+0x94/0xa9

Sep 13 21:11:39 wn122 kernel: [<ffffffff800f4fa9>] compat_sys_getdents+0xaf/0xbd

Sep 13 21:11:39 wn122 kernel: [<ffffffff8000c9df>] do_path_lookup+0x270/0x2e8

Sep 13 21:11:39 wn122 kernel:  [<ffffffff800123f7>] getname+0x15b/0x1c1

Sep 13 21:11:39 wn122 kernel: [<ffffffff80023318>] __user_walk_fd+0x37/0x4c

Sep 13 21:11:39 wn122 kernel: [<ffffffff800320b8>] sys_faccessat+0xe4/0x18d

Sep 13 21:11:39 wn122 kernel:  [<ffffffff80034cd4>] vfs_readdir+0x94/0xa9

Sep 13 21:11:39 wn122 kernel: [<ffffffff800f4fa9>] compat_sys_getdents+0xaf/0xbd

Sep 13 21:11:39 wn122 kernel: [<ffffffff8006149b>] sysenter_do_call+0x1b/0x67

Sep 13 21:11:39 wn122 kernel: [<ffffffff8002c205>] dummy_inode_permission+0x0/0x3

Sep 13 21:11:39 wn122 kernel:

Sep 13 21:11:39 wn122 kernel: LustreError: dumping log to /tmp/lustre-log.1284405099.27016

Sep 13 21:11:44 wn122 dhclient: DHCPREQUEST on eth0 to 148.187.67.113 port 67

Sep 13 21:11:49 wn122 kernel: BUG: soft lockup - CPU#3 stuck for 10s! [ptlrpcd:31817]

Sep 13 21:11:49 wn122 kernel: CPU 3:

Sep 13 21:11:49 wn122 kernel: Modules linked in: mgc(U) lustre(U) lov(U) mdc(U) lquota(U) osc(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) libcfs(U) nfs fscache nfs_acl loc

kd sunrpc bonding(U) ip_conntrack_netbios_ns ipt_REJECT xt_tcpudp xt_state iptable_filter iptable_nat ip_nat ip_conntrack nfnetlink iptable_mangle ip_tables x_tables rdma_ucm(U) ib

_sdp(U) rdma_cm(U) iw_cm(U) ib_addr(U) ib_ipoib(U) ipoib_helper(U) ib_cm(U) ipv6 xfrm_nalgo crypto_api ib_uverbs(U) ib_umad(U) mlx4_vnic(U) ib_sa(U) mlx4_ib(U) ib_mthca(U) ib_mad(U

) ib_core(U) dm_multipath scsi_dh video hwmon backlight sbs i2c_ec button battery asus_acpi acpi_memhotplug ac parport_pc lp parport joydev sg i2c_i801 i2c_core e1000e shpchp mlx4_

core(U) pcspkr dm_raid45 dm_message dm_region_hash dm_mem_cache dm_snapshot dm_zero dm_mirror dm_log dm_mod ata_piix libata sd_mod scsi_mod ext3 jbd uhci_hcd ohci_hcd ehci_hcd

Sep 13 21:11:49 wn122 kernel: Pid: 31817, comm: ptlrpcd Tainted: G 2.6.18-128.7.1.el5 #1

Sep 13 21:11:49 wn122 kernel: RIP: 0010:[<ffffffff80064cb7>] [<ffffffff80064cb7>] .text.lock.spinlock+0x5/0x30

Sep 13 21:11:49 wn122 kernel: RSP: 0018:ffff8101ec177cb8  EFLAGS: 00000282

Sep 13 21:11:49 wn122 kernel: RAX: 000000000000004f RBX: 0000000000000000 RCX: 0000000000000000

Sep 13 21:11:49 wn122 kernel: RDX: ffff81035956b480 RSI: ffff810253c2d400 RDI: ffff810552ccb500

Sep 13 21:11:49 wn122 kernel: RBP: ffff810192294000 R08: 5a5a5a5a5a5a5a5a R09: 5a5a5a5a5a5a5a5a

Sep 13 21:11:49 wn122 kernel: R10: 5a5a5a5a5a5a5a5a R11: 5a5a5a5a5a5a5a5a R12: 0000000000000038

Sep 13 21:11:49 wn122 kernel: R13: ffff81045b0150c0 R14: ffff81067fc57000 R15: ffffffff886f5168

Sep 13 21:11:49 wn122 kernel: FS: 00002b5af649d240(0000) GS:ffff81010c4c8e40(0000) knlGS:0000000000000000

Sep 13 21:11:49 wn122 kernel: CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b

Sep 13 21:11:49 wn122 kernel: CR2: 0000000008183094 CR3: 0000000000201000 CR4: 00000000000006e0

Sep 13 21:11:49 wn122 kernel:

Sep 13 21:11:49 wn122 kernel: Call Trace:

Sep 13 21:11:49 wn122 kernel: [<ffffffff889249dc>] :lustre:ll_statahead_interpret+0xfc/0x5b0

Sep 13 21:11:49 wn122 kernel: [<ffffffff88861779>] :mdc:mdc_intent_getattr_async_interpret+0x459/0x490

Sep 13 21:11:49 wn122 kernel: [<ffffffff88729246>] :ptlrpc:ptlrpc_check_set+0x1186/0x1440

Sep 13 21:11:49 wn122 kernel: [<ffffffff8004ac1b>] try_to_del_timer_sync+0x51/0x5a

Sep 13 21:11:49 wn122 kernel: [<ffffffff8875d05d>] :ptlrpc:ptlrpcd_check+0xdd/0x1f0

Sep 13 21:11:49 wn122 kernel: [<ffffffff80095004>] process_timeout+0x0/0x5

Sep 13 21:11:49 wn122 kernel: [<ffffffff8875d9a1>] :ptlrpc:ptlrpcd+0x1b1/0x259

Sep 13 21:11:49 wn122 kernel: [<ffffffff8008a4b4>] default_wake_function+0x0/0xe

Sep 13 21:11:49 wn122 kernel:  [<ffffffff8005dfb1>] child_rip+0xa/0x11

Sep 13 21:11:49 wn122 kernel: [<ffffffff8875d7f0>] :ptlrpc:ptlrpcd+0x0/0x259

Sep 13 21:11:49 wn122 kernel:  [<ffffffff8005dfa7>] child_rip+0x0/0x11

Sep 13 21:11:49 wn122 kernel:

Sep 13 21:11:49 wn122 kernel: BUG: soft lockup - CPU#0 stuck for 10s! [ll_sa_27016:28593]

Sep 13 21:11:49 wn122 kernel: CPU 0:

Sep 13 21:11:49 wn122 kernel: Modules linked in: mgc(U) lustre(U) lov(U) mdc(U) lquota(U) osc(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) libcfs(U) nfs fscache nfs_acl lockd sunrpc bonding(U) ip_conntrack_netbios_ns ipt_REJECT xt_tcpudp xt_state iptable_filter iptable_nat ip_nat ip_conntrack nfnetlink iptable_mangle ip_tables x_tables rdma_ucm(U) ib_sdp(U) rdma_cm(U) iw_cm(U) ib_addr(U) ib_ipoib(U) ipoib_helper(U) ib_cm(U) ipv6 xfrm_nalgo crypto_api ib_uverbs(U) ib_umad(U) mlx4_vnic(U) ib_sa(U) mlx4_ib(U) ib_mthca(U) ib_mad(U) ib_core(U) dm_multipath scsi_dh video hwmon backlight sbs i2c_ec button battery asus_acpi acpi_memhotplug ac parport_pc lp parport joydev sg i2c_i801 i2c_core e1000e shpchp mlx4_core(U) pcspkr dm_raid45 dm_message dm_region_hash dm_mem_cache dm_snapshot dm_zero dm_mirror dm_log dm_mod ata_piix libata sd_mod scsi_mod ext3 jbd uhci_hcd ohci_hcd ehci_hcd

Sep 13 21:11:49 wn122 kernel: Pid: 28593, comm: ll_sa_27016 Tainted: G 2.6.18-128.7.1.el5 #1

Sep 13 21:11:49 wn122 kernel: RIP: 0010:[<ffffffff80064cb4>] [<ffffffff80064cb4>] .text.lock.spinlock+0x2/0x30

Sep 13 21:11:49 wn122 kernel: RSP: 0000:ffff810251edfcf8  EFLAGS: 00000282

Sep 13 21:11:49 wn122 kernel: RAX: 0000000000000001 RBX: ffff8101cf193e80 RCX: 0000000000000000

Sep 13 21:11:49 wn122 kernel: RDX: 0000000000000012 RSI: 000000000abe79b0 RDI: ffff810552ccb500

Sep 13 21:11:49 wn122 kernel: RBP: 0000000000000286 R08: 0000000351edfde0 R09: 0000000000000000

Sep 13 21:11:49 wn122 kernel: R10: ffff810311d9a400 R11: 0000000000000248 R12: ffff81023f907a00

Sep 13 21:11:49 wn122 kernel: R13: ffffffff88726dba R14: ffff810162543bc0 R15: 0000000000000000

Sep 13 21:11:49 wn122 kernel: FS: 0000000000000000(0000) GS:ffffffff803ac000(0000) knlGS:0000000000000000

Sep 13 21:11:49 wn122 kernel: CS: 0010 DS: 002b ES: 002b CR0: 000000008005003b

Sep 13 21:11:49 wn122 kernel: CR2: 000000000c0f900c CR3: 000000066b863000 CR4: 00000000000006e0

Sep 13 21:11:49 wn122 kernel:

Sep 13 21:11:49 wn122 kernel: Call Trace:

Sep 13 21:11:49 wn122 kernel: [<ffffffff88923787>] :lustre:ll_sai_entry_to_stated+0x87/0x330

Sep 13 21:11:49 wn122 kernel:  [<ffffffff8000d0de>] dput+0x2c/0x114

Sep 13 21:11:49 wn122 kernel: [<ffffffff88923f8b>] :lustre:do_statahead_interpret+0x55b/0x5c0

Sep 13 21:11:49 wn122 kernel: [<ffffffff88925641>] :lustre:ll_statahead_thread+0x7b1/0x1750

Sep 13 21:11:49 wn122 kernel: [<ffffffff8008a4b4>] default_wake_function+0x0/0xe

Sep 13 21:11:49 wn122 kernel:  [<ffffffff8005dfb1>] child_rip+0xa/0x11

Sep 13 21:11:49 wn122 kernel: [<ffffffff88924e90>] :lustre:ll_statahead_thread+0x0/0x1750

Sep 13 21:11:49 wn122 kernel:  [<ffffffff8005dfa7>] child_rip+0x0/0x11

After we had 3 clients go into this state the first day, and 2 the next, we decided to downgrade the clients back to 1.8.3, while leaving the servers at 1.8.4. This morning, we had 2 more clients get stuck again, so my assumption is that there is a problem with the 1.8.4 servers.

This ASSERTION looks related with statahead, such feature only affects client-side behavior, but nothing for server. On the other hand, we never saw that on lustre-1.8.3 or former version, and as I known, there were almost no changes for statahead between lustre-1.8.3 and lustre-1.8.4. So I am not sure whether your assumption is right not.

Anyway, to make your system run, you can disable statahead on client-side firstly, to check what will happen after that. The client-side proc interface "statahead_max" is for that.

Cheers,
Nasf

These are RedHat 5.4 servers and clients, with the 2.6.18-194.3.1.el5_lustre.1.8.4 kernel.

Any help on this issue would be great.

Thanks,

Jason Temple


_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss

_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss

Reply via email to