Hi all,

this is a follow up some help I was soliciting with my corrupted pool.

The short story is I can have no confidence in the quality in the labels on 2
of my 5 drive RAIDZ array.  For various reasons.

There is a possibility even that one drive has label of another (a mirroring
accident).

Anyhoo, for some odd reason, the drives finally mounted (they are actually
drive images on another ZFS pool which I have snapshotted).

When I imported the pool, ZFS complained that two of the datasets would not
mount, but the remainder did.

It seems that small files read ok.  (Perhaps small enough to fit on a single 
block -
hence probably mirrored and not striped.  Assuming my understanding of what
happens to small files is correct).

But on larger files I get:

root@openindiana-01:/ZP-8T-RZ1-01/incoming# cp httpd-error.log.zip /mnt2/
cp: reading `httpd-error.log.zip': I/O error

and on some directories:

root@openindiana-01:/ZP-8T-RZ1-01/usr# ls -al
cd ..ls: cannot access obj: Bad exchange descriptor
total 54
drwxr-xr-x  5 root root  5 2011-11-03 16:28 .
drwxr-xr-x 11 root root 11 2011-11-04 13:14 ..
??????????  ? ?    ?     ?                ? obj
drwxr-xr-x 68 root root 83 2011-10-30 01:00 ports
drwxr-xr-x 22 root root 31 2011-09-25 02:00 src

Here is the zpool status output:

root@openindiana-01:/ZP-8T-RZ1-01# zpool status
 pool: ZP-8T-RZ1-01
state: DEGRADED
status: One or more devices has experienced an error resulting in data
       corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
       entire pool from backup.
  see: http://www.sun.com/msg/ZFS-8000-8A
 scan: scrub in progress since Sat Nov  5 23:57:46 2011
   112G scanned out of 6.93T at 6.24M/s, 318h17m to go
   305M repaired, 1.57% done
config:

       NAME                      STATE     READ WRITE CKSUM
       ZP-8T-RZ1-01              DEGRADED     0     0  356K
         raidz1-0                DEGRADED     0     0  722K
           12339070507640025002  UNAVAIL      0     0     0  was /dev/lofi/2
           /dev/lofi/5           DEGRADED     0     0     0  too many errors 
(repairing)
           /dev/lofi/4           DEGRADED     0     0     0  too many errors 
(repairing)
           /dev/lofi/3           DEGRADED     0     0 74.4K  too many errors 
(repairing)
           /dev/lofi/1           DEGRADED     0     0     0  too many errors 
(repairing)

All those errors may be caused by one disk actually owning the wrong label.
I'm not entirely sure.

Also, while it's complaining that /dev/lofi/2 is UNAVAIL, it certainly is.
Although it's probably not labelled with '12339070507640025002'.

I'd love to get some of my data back.  Any recovery is a bonus.

If anyone is keen, I have enabled SSH into the Open Indiana box
which I'm using to try and recovery the pool, so if you'd like to take a shot
please let me know.

Thanks in advance,
Scott
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to