Hello ZFS wizards, Have an odd ZFS problem I'd like to run by you - Root pool on this machine is a 'simple' mirror - just two disks. # zpool status NAME STATE READ WRITE CKSUM rpool ONLINE 0 0 3 mirror-0 ONLINE 0 0 6 c2t0d0s0 ONLINE 0 0 6 c2t1d0s0 ONLINE 0 0 6 errors: Permanent errors have been detected in the following files: rpool/ROOT/openindiana-userland-154@zfs-auto-snap_monthly-2011-11-22-09h19:/etc/svc/repository-boot-tmpEdaGba ... or similar; CKSUM counts have varied, but were always in that 1x - 2x , 'symmetrical' pattern. After working through the problems above, scrubbing and zfs destroying the snapshot with 'permanent errors', the CKSUMS clear up, but vestiges of the file remain as hex addresses: NAME STATE READ WRITE CKSUM rpool ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 c2t0d0s0 ONLINE 0 0 0 c2t1d0s0 ONLINE 0 0 0 errors: Permanent errors have been detected in the following files: <0x18e73>:<0x78007> I have no evidence that ZFS is itself the direct culprit here; it may just be on the receiving end of one of the couple of problems we've recently worked through on this machine: 1. a defective CPU, managed by the fault manager, but without a fully-configured crashdump (now rectified), then 2. the SandyBridge 'interrupt storm' problem, which we seem to have now worked around. The storage pools are scrubbed pretty regularly, and we generally have no cksum errors at all. At one point, vmstat reported 7+ _million+ interrupt faults over 5 seconds! I've attempted to clear stats on the pool as well (didn't expect this to work, but worth a try, right?) Important to note that Memtest+ had been run, last time for ~14 hrs, with no error reported. Don't think the storage controller is the culprit, either, as _all_ drives are controlled by the P67A - and no other problems seen. And no errors reported via smartctl. Would welcome input from two perspectives: 1) Before I rebuild the pool/reinstall/whatever, is anyone here interested in any diagnostic output which might still be available? Is any of this useful as a bug report? 2) Then, would love to hear ideas on a solution. Proposed solutions include: 1) creating new BE based on snap of root pool: - Snapshot root pool - (zfs send to datapool for safekeeping) - Split rpool - zpool create newpool (on Drive 'B') - beadm -p create newpool NEWboot (being sure to use slice 0 of Drive 'B') 2) Simply deleting _all_ snapshots on the rpool. 3) complete re-install Tks for feedback. Lou Picciano
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss