Weird. I have no idea how you could remove that file (beside destroying the 
entire filesystem)...

One other thing I noticed:

        NAME        STATE     READ WRITE CKSUM
        rpool       ONLINE       0     0     8
          raidz1    ONLINE       0     0     8
            c0t7d0  ONLINE       0     0     0
            c0t1d0  ONLINE       0     0     0
            c0t2d0  ONLINE       0     0     0

When you see non-zero CKSUM error counters at the pool or raidz1/z2 vdev 
level, but no error on the devices like this, it means that ZFS couldn't 
correct the corruption errors after multiple attempts of reconstructing the 
stripes, each time assuming a different device was corrupting data. IOW it 
means that 2+ (in a raidz1) or 3+ (in a raidz2) devices returned corrupted 
data in the same stripe. Since it is statistically improbable to have that 
many silent data corruption in the same stripe, most likely this condition 
indicates a hardware problem. I suggest running memtest to stress-test your 
cpu/mem/mobo.

-marc

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to