Today I encountered data corruption on two zfs pools due to a RAM failure in my
OI box running on a dell T710. My rpool now looks like this (after reboot):
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
scan: scrub repaired 0 in 1h1m with 1 errors on Tue Jan 31 19:59:50 2012
NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 1
mirror-0 ONLINE 0 0 2
c4t50014EE10313DE5Dd0s0 ONLINE 0 0 2
c4t50014EE158688073d0s0 ONLINE 0 0 2
errors: Permanent errors have been detected in the following files:
I have 17 files that are permanently corrupted. The corruption of gdm/core was
found while scrubbing the pool. All the other 16 files where displayed as
corrupted after the pool fell in degraded state. I'm not sure if these files
are really corrupted, though: I can access all these files and e.g.
/usr/gnu/bin/rm works with no faults. All files have the identical md5 sum
compared the the corresponding files of a different box, also running the same
version of OI.
How do I find out, if these files are corrupted? If they appear to be ok, how
do I get rid of the errors?
How can two healthy pools get that messed up, when a RAM DIMM gets broken?
zfs-discuss mailing list