Hello Eric, Monday, June 12, 2006, 11:21:24 PM, you wrote:
ES> I reproduced this pretty easily on a lab machine. I've filed: ES> 6437568 ditto block repair is incorrectly propagated to root vdev ES> To track this issue. Keep in mind that you do have a flakey ES> controller/lun/something. If this had been a user data block, your data ES> would be gone. I belive that something else is also happening here. I can see CKSUM errors on two different servers (v240 and T2000) all on non-redundant zpools and all the times it looks like ditto block helped - hey, it's just improbable. And while on T2000 from fmdump -ev I get: Jul 05 19:59:43.8786 ereport.io.fire.pec.btp 0x14e4b8015f612002 Jul 05 20:05:28.9165 ereport.io.fire.pec.re 0x14e5f951ce12b002 Jul 05 20:05:58.5381 ereport.io.fire.pec.re 0x14e614e78f4c9002 Jul 05 20:05:58.5389 ereport.io.fire.pec.btp 0x14e614e7b6ddf002 Jul 05 23:34:11.1960 ereport.io.fire.pec.re 0x1513869a6f7a6002 Jul 05 23:34:11.1967 ereport.io.fire.pec.btp 0x1513869a95196002 Jul 06 00:09:17.1845 ereport.io.fire.pec.re 0x151b2fca4c988002 Jul 06 00:09:17.1852 ereport.io.fire.pec.btp 0x151b2fca72e6b002 on v240 fmdump shows nothing for over a month and I'm sure I did zpool clear on that server later. v240: bash-3.00# zpool status nfs-s5-s7 pool: nfs-s5-s7 state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: http://www.sun.com/msg/ZFS-8000-9P scrub: none requested config: NAME STATE READ WRITE CKSUM nfs-s5-s7 ONLINE 0 0 167 c4t600C0FF00000000009258F28706F5201d0 ONLINE 0 0 167 errors: No known data errors bash-3.00# bash-3.00# zpool clear nfs-s5-s7 bash-3.00# zpool status nfs-s5-s7 pool: nfs-s5-s7 state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM nfs-s5-s7 ONLINE 0 0 0 c4t600C0FF00000000009258F28706F5201d0 ONLINE 0 0 0 errors: No known data errors bash-3.00# bash-3.00# zpool scrub nfs-s5-s7 bash-3.00# zpool status nfs-s5-s7 pool: nfs-s5-s7 state: ONLINE scrub: scrub in progress, 0.01% done, 269h24m to go config: NAME STATE READ WRITE CKSUM nfs-s5-s7 ONLINE 0 0 0 c4t600C0FF00000000009258F28706F5201d0 ONLINE 0 0 0 errors: No known data errors bash-3.00# We'll see the result - I hope I would have not to stop it in the morning. Anyway I have a feeling that nothing will be reported. ps. I've got several similar pools on those two servers and I see CKSUM errors on all of them with the same result - it's almost impossible. -- Best regards, Robert mailto:[EMAIL PROTECTED] http://milek.blogspot.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss