[zfs-discuss] Pool error in a hex file name

2011-06-13 Thread James Sutherland
Anyone know what this means?  After a scrub I apparently have an error in a 
file name that I don't understand:

zpool status -v pumbaa1
  pool: pumbaa1
 state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
pumbaa1 ONLINE   0 0 0
  c0d0s0ONLINE   0 0 0

errors: Permanent errors have been detected in the following files:

0x4e7:0x37191
$
-
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Pool error in a hex file name

2011-06-13 Thread James Sutherland
A reboot and then another scrub fixed this.  Reboot made no difference.  So 
after the reboot I started another scrub and now the pool shows clean.  

So the sequence was like this:
1. zpool reported ioerrors after a scrub with an error on a file in a snapshot
2. destroyed the snapshot with the error
3. zpool errror changed to hex addressed instead of filename and snapshot name
4. scrubbed pool again and error went away.

All this happened after having a disk go bad.  Started getting lots of hard 
errors.  My ZFS monitor script e-mailed me and let me know.  So I added another 
disk.  Mirrored the errored pool.  Removed the failing disk and removed that 
disk from the pool too. Now running the pool on just the new disk.

 I'm out of ports on this hardware so I can't have a permanent mirror on this 
pool.  This pool is only the boot pool anyway so if it where to be lost I 
wouldn't loose much.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Pool error in a hex file name

2011-06-13 Thread Richard Elling
On Jun 12, 2011, at 1:53 PM, James Sutherland wrote:

 A reboot and then another scrub fixed this.  Reboot made no difference.  So 
 after the reboot I started another scrub and now the pool shows clean.  
 
 So the sequence was like this:
 1. zpool reported ioerrors after a scrub with an error on a file in a snapshot
 2. destroyed the snapshot with the error
 3. zpool errror changed to hex addressed instead of filename and snapshot name
 4. scrubbed pool again and error went away.

zpool status shows the current errors and the last errors, so you have to clear 
twice
to make them go away completely. After you destroyed the data, ZFS knew there 
was
bad data, but could no longer tie it to a file name, hence the hex addresses.

 
 All this happened after having a disk go bad.  Started getting lots of hard 
 errors.  My ZFS monitor script e-mailed me and let me know.  So I added 
 another disk.  Mirrored the errored pool.  Removed the failing disk and 
 removed that disk from the pool too. Now running the pool on just the new 
 disk.
 
 I'm out of ports on this hardware so I can't have a permanent mirror on this 
 pool.  This pool is only the boot pool anyway so if it where to be lost I 
 wouldn't loose much.

 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss