I've run into this problem twice now, before I had 10x500GB drives in a ZFS+ 
setup and now again in a 12x500GB ZFS+ setup.

The problem is when the pool reaches ~85% capacity I get random read failures 
and around ~90% capacity I get read failures AND zpool corruption.  For example:

-I open a directory that I know for a fact has files and folders in it but it 
either shows 0 items or hangs on a directory listing
-I try to copy a file from the zpool volume to another volume and it hangs then 
fails

In both these situations if I do a 'zpool status' after the fact it claims that 
the volume has experienced an unrecoverable error and I should find the faulty 
drive and replace it blah blah.  If I do a 'zpool scrub' it eventually reports 
0 faults or error, also if I restart the machine it will usually work jut fine 
again (ie I can read the directory and copy files again).

Is this a systemic problem at 90% capacity or do I perhaps have a faulty drive 
in the array that only gets hit at 90%?  If it is a faulty drive why does 
'zpool status' report fully good health, that makes it hard to find the problem 
drive?

Thanks,
Sam
-- 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to