I've run into this problem twice now, before I had 10x500GB drives in a ZFS+ setup and now again in a 12x500GB ZFS+ setup.
The problem is when the pool reaches ~85% capacity I get random read failures and around ~90% capacity I get read failures AND zpool corruption. For example: -I open a directory that I know for a fact has files and folders in it but it either shows 0 items or hangs on a directory listing -I try to copy a file from the zpool volume to another volume and it hangs then fails In both these situations if I do a 'zpool status' after the fact it claims that the volume has experienced an unrecoverable error and I should find the faulty drive and replace it blah blah. If I do a 'zpool scrub' it eventually reports 0 faults or error, also if I restart the machine it will usually work jut fine again (ie I can read the directory and copy files again). Is this a systemic problem at 90% capacity or do I perhaps have a faulty drive in the array that only gets hit at 90%? If it is a faulty drive why does 'zpool status' report fully good health, that makes it hard to find the problem drive? Thanks, Sam -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss