Hi,
While testing 1.5.97 on kernel 2.6.18, I had the following Oops on a client. Has this been seen before? *Unable to handle kernel paging request at ffff88000b17fff8 RIP: * [<ffffffff8843951a>] :osc:osc_brw_prep_request+0x3e3/0x9e2 PGD 1675067 PUD 1676067 PMD 16cf067 PTE 0 Oops: 0000 [1] SMP CPU 0 Modules linked in: xt_tcpudp xt_physdev iptable_filter ip_tables x_tables osc mgc lustre lov lquota mdc ksocklnd ptlrpc obdclass lnet lvfs libcfs bridge netloop ipv6 dm_snapshot dm_mirror dm_mod usbhid usbkbd ipmi_watchdog ipmi_devintf ipmi_poweroff ipmi_si ipmi_msghandler dummy loop ide_generic ide_disk evdev psmouse shpchp pci_hotplug pcspkr serio_raw i2c_piix4 i2c_core ext3 jbd mbcache sd_mod ide_cd cdrom sata_svw libata scsi_mod tg3 ehci_hcd generic ohci_hcd serverworks ide_core fan Pid: 3638, comm: cat Not tainted 2.6.18-3-xen-amd64 #1 RIP: e030:[<ffffffff8843951a>] [<ffffffff8843951a>] :osc:osc_brw_prep_request+0x3e3/0x9e2 RSP: e02b:ffff88000fbf1798 EFLAGS: 00010287 RAX: 0000000000000001 RBX: ffff88000eb79ce0 RCX: 0000000000000000 RDX: ffff88000b180000 RSI: ffff88000b180000 RDI: ffff880026c53cb0 RBP: ffff88005c6e4170 R08: 0000000000000000 R09: 0000000000000000 R10: 0000000000000000 R11: 0000000000000000 R12: ffff88005c6e4088 R13: ffff88005c701a00 R14: ffff88005c58e498 R15: 0000000000000000 FS: 00002b0e147c86d0(0000) GS:ffffffff804c3000(0000) knlGS:0000000000000000 CS: e033 DS: 0000 ES: 0000 Process cat (pid: 3638, threadinfo ffff88000fbf0000, task ffff8800626cc7f0) Stack: ffff88000b180000 00000100802a6c2b ffff880026c53cb0 ffff88005c58e498 0000000000000001 0000000000000010 0000001062948240 ffff880056946000 0000000102a13e28 0000000400000000 Call Trace: [<ffffffff8843d611>] :osc:osc_send_oap_rpc+0x80a/0xe6d [<ffffffff8843ddc0>] :osc:osc_check_rpcs+0x14c/0x29b [<ffffffff88445751>] :osc:osc_queue_async_io+0xc2b/0xd0a [<ffffffff8820e41e>] :libcfs:libcfs_debug_vmsg2+0x600/0x897 [<ffffffff883a0d38>] :lov:lov_queue_async_io+0x2f1/0x3a1 [<ffffffff883f996f>] :lustre:queue_or_sync_write+0x2b6/0xc43 [<ffffffff883fcc99>] :lustre:ll_commit_write+0x269/0x5df [<ffffffff80210ba1>] generic_file_buffered_write+0x438/0x646 [<ffffffff883b742f>] :lov:lov_update_enqueue_set+0x345/0x3a1 [<ffffffff8020f13c>] current_fs_time+0x3b/0x40 [<ffffffff884424b5>] :osc:osc_enqueue+0xfb/0x48e [<ffffffff8021646f>] __generic_file_aio_write_nolock+0x2e4/0x32f [<ffffffff883ef54b>] :lustre:ll_inode_size_unlock+0x81/0xd6 [<ffffffff802a2809>] __generic_file_write_nolock+0x8f/0xa8 [<ffffffff882e4886>] :ptlrpc:ldlm_completion_ast+0x0/0x5e2 [<ffffffff80290415>] autoremove_wake_function+0x0/0x2e [<ffffffff88408367>] :lustre:lt_get_mmap_locks+0x2b1/0x3a5 [<ffffffff8024529d>] generic_file_write+0x49/0xa7 [<ffffffff883ea8f3>] :lustre:ll_file_write+0x669/0x7e7 [<ffffffff80216b9b>] vfs_write+0xce/0x174 [<ffffffff802173be>] sys_write+0x45/0x6e [<ffffffff8025c81a>] system_call+0x86/0x8b [<ffffffff8025c794>] system_call+0x0/0x8b Note: the client was running on a Xen dom0, with fairly small memory, so lustre is not necessarily at fault. Further details on request. - Alastair McKinstry
_______________________________________________ Lustre-discuss mailing list [email protected] https://mail.clusterfs.com/mailman/listinfo/lustre-discuss
