Today we had another panic, at least it was during work time :) Just a 
shame the 999GB ufs takes 80+ mins to fsck. (Yes, it is mounted 'logging').







panic[cpu3]/thread=ffffff001e70dc80:
free: freeing free block, dev:0xb600000024, block:13144, ino:1737885, 
fs:/export
/saba1


ffffff001e70d500 genunix:vcmn_err+28 ()
ffffff001e70d550 ufs:real_panic_v+f7 ()
ffffff001e70d5b0 ufs:ufs_fault_v+1d0 ()
ffffff001e70d6a0 ufs:ufs_fault+a0 ()
ffffff001e70d770 ufs:free+38f ()
ffffff001e70d830 ufs:indirtrunc+260 ()
ffffff001e70dab0 ufs:ufs_itrunc+738 ()
ffffff001e70db60 ufs:ufs_trans_itrunc+128 ()
ffffff001e70dbf0 ufs:ufs_delete+3b0 ()
ffffff001e70dc60 ufs:ufs_thread_delete+da ()
ffffff001e70dc70 unix:thread_start+8 ()

syncing file systems...

panic[cpu3]/thread=ffffff001e70dc80:
panic sync timeout

dumping to /dev/dsk/c6t0d0s1, offset 65536, content: kernel


 > $c
vpanic()
vcmn_err+0x28(3, fffffffff783a128, ffffff001e70d678)
real_panic_v+0xf7(0, fffffffff783a128, ffffff001e70d678)
ufs_fault_v+0x1d0(ffffff04facf65c0, fffffffff783a128, ffffff001e70d678)
ufs_fault+0xa0()
free+0x38f(ffffff001e70d8d0, a6a7358, 2000, 89)
indirtrunc+0x260(ffffff001e70d8d0, a6a42b8, ffffffffffffffff, 0, 89)
ufs_itrunc+0x738(ffffff0550b9fde0, 0, 81, fffffffec0594db0)
ufs_trans_itrunc+0x128(ffffff0550b9fde0, 0, 81, fffffffec0594db0)
ufs_delete+0x3b0(fffffffed20e2a00, ffffff0550b9fde0, 1)
ufs_thread_delete+0xda(ffffffff64704840)
thread_start+8()

 > ::panicinfo
              cpu                3
           thread ffffff001e70dc80
          message
free: freeing free block, dev:0xb600000024, block:13144, ino:1737885, 
fs:/export
/saba1
              rdi fffffffff783a128
              rsi ffffff001e70d678
              rdx fffffffff783a128
              rcx ffffff001e70d678
               r8 fffffffff783a128
               r9                0
              rax                3
              rbx                0
              rbp ffffff001e70d4d0
              r10 fffffffec3d40580
              r10 fffffffec3d40580
              r11 ffffff001e70dc80
              r12 fffffffff783a128
              r13 ffffff001e70d678
              r14                3
              r15 fffffffff783a128
           fsbase                0
           gsbase fffffffec3d40580
               ds               4b
               es               4b
               fs                0
               gs              1c3
           trapno                0
              err                0
              rip fffffffffb83c860
               cs               30
           rflags              246
              rsp ffffff001e70d488
               ss               38
           gdt_hi                0
           gdt_lo         800001ef
           idt_hi                0
           idt_lo         70000fff
              ldt                0
             task               70
              cr0         8005003b
              cr2         fed0e010
              cr3          2c00000
              cr4              6f8





Jorgen Lundman wrote:
> On Saturday the X4500 system paniced, and rebooted. For some reason the 
> /export/saba1 UFS partition was corrupt, and needed "fsck". This is why 
> it did not come back online. /export/saba1 is mounted "logging,noatime", 
> so fsck should never (-ish) be needed.
> 
> SunOS x4500-01.unix 5.11 snv_70b i86pc i386 i86pc
> 
> /export/saba1 on /dev/zvol/dsk/zpool1/saba1 
> read/write/setuid/devices/intr/largefiles/logging/quota/xattr/noatime/onerror=panic/dev=2d80024
>  
> on Sat Jul  5 08:48:54 2008
> 
> 
> One possible related bug:
> 
> http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=4884138
> 
> 
> What would be the best solution? Go back to latest Solaris 10 and pass 
> it on to Sun support, or find a patch for this problem?
> 
> 
> 
> Panic dump follows:
> 
> 
> -rw-r--r--   1 root     root     2529300 Jul  5 08:48 unix.2
> -rw-r--r--   1 root     root     10133225472 Jul  5 09:10 vmcore.2
> 
> 
> # mdb unix.2 vmcore.2
> Loading modules: [ unix genunix specfs dtrace cpu.AuthenticAMD.15 uppc 
> pcplusmp scsi_vhci ufs md ip hook neti sctp arp usba uhci s1394 qlc fctl 
> nca lofs zfs random cpc crypto fcip fcp logindmux nsctl sdbc ptm sv ii 
> sppp rdc nfs ]
> 
>  > $c
> vpanic()
> vcmn_err+0x28(3, fffffffff783ade0, ffffff001e737aa8)
> real_panic_v+0xf7(0, fffffffff783ade0, ffffff001e737aa8)
> ufs_fault_v+0x1d0(fffffffed0bfb980, fffffffff783ade0, ffffff001e737aa8)
> ufs_fault+0xa0()
> dqput+0xce(ffffffff1db26ef0)
> dqrele+0x48(ffffffff1db26ef0)
> ufs_trans_dqrele+0x6f(ffffffff1db26ef0)
> ufs_idle_free+0x16d(ffffff04f17b1e00)
> ufs_idle_some+0x152(3f60)
> ufs_thread_idle+0x1a1()
> thread_start+8()
> 
> 
>  > ::cpuinfo
>   ID ADDR             FLG NRUN BSPL PRI RNRN KRNRN SWITCH THREAD 
>     PROC
>    0 fffffffffbc2fc10  1b    0    0  60   no    no t-0 
> ffffff001e737c80 sched
>    1 fffffffec3a0a000  1f    1    0  -1   no    no t-0    ffffff001e971c80
>   (idle)
>    2 fffffffec3a02ac0  1f    0    0  -1   no    no t-1    ffffff001e9dbc80
>   (idle)
>    3 fffffffec3d60580  1f    0    0  -1   no    no t-1    ffffff001ea50c80
>   (idle)
> 
>  > ::panicinfo
>               cpu                0
>            thread ffffff001e737c80
>           message dqput: dqp->dq_cnt == 0
>               rdi fffffffff783ade0
>               rsi ffffff001e737aa8
>               rdx fffffffff783ade0
>               rcx ffffff001e737aa8
>                r8 fffffffff783ade0
>                r9                0
>               rax                3
>               rbx                0
>               rbp ffffff001e737900
>               r10 fffffffffbc26fb0
>               r10 fffffffffbc26fb0
>               r11 ffffff001e737c80
>               r12 fffffffff783ade0
>               r13 ffffff001e737aa8
>               r14                3
>               r15 fffffffff783ade0
>            fsbase                0
>            gsbase fffffffffbc26fb0
>                ds               4b
>                es               4b
>            fsbase                0
>            gsbase fffffffffbc26fb0
>                ds               4b
>                es               4b
>                fs                0
>                gs              1c3
>            trapno                0
>               err                0
>               rip fffffffffb83c860
>                cs               30
>            rflags              246
>               rsp ffffff001e7378b8
>                ss               38
>            gdt_hi                0
>            gdt_lo         e00001ef
>            idt_hi                0
>            idt_lo         77c00fff
>               ldt                0
>              task               70
>               cr0         8005003b
>               cr2         fee7d650
>               cr3          2c00000
>               cr4              6f8
> 
>  > ::msgbuf
> quota_ufs: over hard disk limit (pid 600, uid 178199, inum 941499, fs 
> /export/zero1)
> quota_ufs: over hard disk limit (pid 600, uid 33647, inum 29504134, fs 
> /export/zero1)
> 
> panic[cpu0]/thread=ffffff001e737c80:
> dqput: dqp->dq_cnt == 0
> 
> 
> ffffff001e737930 genunix:vcmn_err+28 ()
> ffffff001e737980 ufs:real_panic_v+f7 ()
> ffffff001e7379e0 ufs:ufs_fault_v+1d0 ()
> ffffff001e737ad0 ufs:ufs_fault+a0 ()
> ffffff001e737b00 ufs:dqput+ce ()
> ffffff001e737b30 ufs:dqrele+48 ()
> ffffff001e737b70 ufs:ufs_trans_dqrele+6f ()
> ffffff001e737bc0 ufs:ufs_idle_free+16d ()
> ffffff001e737c10 ufs:ufs_idle_some+152 ()
> ffffff001e737c60 ufs:ufs_thread_idle+1a1 ()
> ffffff001e737c70 unix:thread_start+8 ()
> 
> syncing file systems...
> 
> 
> 
> 

-- 
Jorgen Lundman       | <[EMAIL PROTECTED]>
Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
Shibuya-ku, Tokyo    | +81 (0)90-5578-8500          (cell)
Japan                | +81 (0)3 -3375-1767          (home)
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to