Hello Eric,

Monday, June 12, 2006, 11:21:24 PM, you wrote:

ES> I reproduced this pretty easily on a lab machine.  I've filed:

ES> 6437568 ditto block repair is incorrectly propagated to root vdev

ES> To track this issue.  Keep in mind that you do have a flakey
ES> controller/lun/something.  If this had been a user data block, your data
ES> would be gone.


I belive that something else is also happening here.
I can see CKSUM errors on two different servers (v240 and T2000) all
on non-redundant zpools and all the times it looks like ditto block
helped - hey, it's just improbable.

And while on T2000 from fmdump -ev I get:

Jul 05 19:59:43.8786 ereport.io.fire.pec.btp               0x14e4b8015f612002
Jul 05 20:05:28.9165 ereport.io.fire.pec.re                0x14e5f951ce12b002
Jul 05 20:05:58.5381 ereport.io.fire.pec.re                0x14e614e78f4c9002
Jul 05 20:05:58.5389 ereport.io.fire.pec.btp               0x14e614e7b6ddf002
Jul 05 23:34:11.1960 ereport.io.fire.pec.re                0x1513869a6f7a6002
Jul 05 23:34:11.1967 ereport.io.fire.pec.btp               0x1513869a95196002
Jul 06 00:09:17.1845 ereport.io.fire.pec.re                0x151b2fca4c988002
Jul 06 00:09:17.1852 ereport.io.fire.pec.btp               0x151b2fca72e6b002


on v240 fmdump shows nothing for over a month and I'm sure I did zpool
clear on that server later.


v240:
bash-3.00# zpool status nfs-s5-s7
  pool: nfs-s5-s7
 state: ONLINE
status: One or more devices has experienced an unrecoverable error.  An
        attempt was made to correct the error.  Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
        using 'zpool clear' or replace the device with 'zpool replace'.
   see: http://www.sun.com/msg/ZFS-8000-9P
 scrub: none requested
config:

        NAME                                     STATE     READ WRITE CKSUM
        nfs-s5-s7                                ONLINE       0     0   167
          c4t600C0FF00000000009258F28706F5201d0  ONLINE       0     0   167

errors: No known data errors
bash-3.00#
bash-3.00# zpool clear nfs-s5-s7
bash-3.00# zpool status nfs-s5-s7
  pool: nfs-s5-s7
 state: ONLINE
 scrub: none requested
config:

        NAME                                     STATE     READ WRITE CKSUM
        nfs-s5-s7                                ONLINE       0     0     0
          c4t600C0FF00000000009258F28706F5201d0  ONLINE       0     0     0

errors: No known data errors
bash-3.00#
bash-3.00# zpool scrub nfs-s5-s7
bash-3.00# zpool status nfs-s5-s7
  pool: nfs-s5-s7
 state: ONLINE
 scrub: scrub in progress, 0.01% done, 269h24m to go
config:

        NAME                                     STATE     READ WRITE CKSUM
        nfs-s5-s7                                ONLINE       0     0     0
          c4t600C0FF00000000009258F28706F5201d0  ONLINE       0     0     0

errors: No known data errors
bash-3.00#

We'll see the result - I hope I would have not to stop it in the
morning. Anyway I have a feeling that nothing will be reported.


ps. I've got several similar pools on those two servers and I see
CKSUM errors on all of them with the same result - it's almost
impossible.

-- 
Best regards,
 Robert                            mailto:[EMAIL PROTECTED]
                                       http://milek.blogspot.com

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to