Hello,

I've been getting warnings that my zfs pool is degraded. At first it was 
complaining about a few corrupt files, which were listed as hex numbers instead 
of filenames, i.e. 

VOL1:<0x0>

After a scrub, a couple of the filenames appeared - turns out they were in 
snapshots I don't really need, so I destroyed those snapshots and started a new 
scrub. Subsequently, I typed " zpool status -v VOL1" ... and the machine 
rebooted. When I could log on again, I looked at /var/log/messages, but found 
nothing interesting prior to the reboot. I typed " zpool status -v VOL1" again, 
whereupon the machine rebooted. When the machine was back up, I stopped the 
scrub, waited a while, then typed "zpool status -v VOL1" again, and this time 
got:


r...@nexenta1:~# zpool status -v VOL1
  pool: VOL1
 state: DEGRADED
status: One or more devices has experienced an unrecoverable error.  An
        attempt was made to correct the error.  Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
        using 'zpool clear' or replace the device with 'zpool replace'.
   see: http://www.sun.com/msg/ZFS-8000-9P
 scan: scrub canceled on Wed Aug 11 11:03:15 2010
config:

        NAME        STATE     READ WRITE CKSUM
        VOL1        DEGRADED     0     0     0
          raidz1-0  DEGRADED     0     0     0
            c2d0    DEGRADED     0     0     0  too many errors
            c3d0    DEGRADED     0     0     0  too many errors
            c4d0    DEGRADED     0     0     0  too many errors
            c5d0    DEGRADED     0     0     0  too many errors

So, I have the following questions:

1) How do I find out which file is corrupt, when I only get something like 
"VOL1:<0x0>" 
2) What could be causing these reboots?
3) How can I fix my pool?

Thanks!
-- 
This message posted from opensolaris.org
_______________________________________________
opensolaris-help mailing list
opensolaris-help@opensolaris.org

Reply via email to