I was following the instructions in ZFS Troubleshooting Guide on how to
replace a disk in the root pool on x86 system. I'm using OpenIndiana, ZFS
pool v.28 with mirrored system rpool. The replacement disk is brand new.
root:~# zpool status
status: One or more devices has experienced an unrecoverable error. An
attempt was made to correct the error. Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
using 'zpool clear' or replace the device with 'zpool replace'.
scan: resilvered 17.6M in 0h0m with 0 errors on Wed Apr 11 17:45:16 2012
NAME STATE READ WRITE CKSUM
rpool DEGRADED 0 0 0
mirror-0 DEGRADED 0 0 0
c2t5000CCA369C55DB8d0s0 OFFLINE 0 126 0
c2t5000CCA369D5231Cd0s0 ONLINE 0 0 0
errors: No known data errors
I'm not very familiar with Solaris partitions and slices so somewhere in
the format/partition commands I must to have made a mistake because when I
try to replace the disk I'm getting the following error:
root:~# zpool replace rpool c2t5000CCA369C55DB8d0s0 c2t5000CCA369C89636d0s0
invalid vdev specification
use '-f' to override the following errors:
/dev/dsk/c2t5000CCA369C89636d0s0 overlaps with
I used -f and it worked but I was wondering is there a way to completely
"reset" the new disk? Remove all partitions and start from scratch.
zfs-discuss mailing list