Michael Stalnaker wrote: > I have a 24 disk SATA array running on Open Solaris Nevada, b78. We had > a drive fail, and I’ve replaced the device but can’t get the system to > recognize that I replaced the drive. > > Zpool status –v shows the failed drive: > > [EMAIL PROTECTED] ~]$ zpool status -v > pool: LogData > state: DEGRADED > status: One or more devices are faulted in response to persistent errors. > Sufficient replicas exist for the pool to continue functioning in a > degraded state. > action: Replace the faulted device, or use 'zpool clear' to mark the device > repaired. > scrub: resilver completed with 0 errors on Wed Feb 27 11:51:45 2008 > config: > > NAME STATE READ WRITE CKSUM > LogData DEGRADED 0 0 0 > raidz2 DEGRADED 0 0 0 > c0t12d0 ONLINE 0 0 0 > c0t5d0 ONLINE 0 0 0 > c0t0d0 ONLINE 0 0 0 > c0t4d0 ONLINE 0 0 0 > c0t8d0 ONLINE 0 0 0 > c0t16d0 ONLINE 0 0 0 > c0t20d0 ONLINE 0 0 0 > c0t1d0 ONLINE 0 0 0 > c0t9d0 ONLINE 0 0 0 > c0t13d0 ONLINE 0 0 0 > c0t17d0 ONLINE 0 0 0 > c0t20d0 FAULTED 0 0 0 too many errors > c0t2d0 ONLINE 0 0 0 > c0t6d0 ONLINE 0 0 0 > c0t10d0 ONLINE 0 0 0 > c0t14d0 ONLINE 0 0 0 > c0t18d0 ONLINE 0 0 0 > c0t22d0 ONLINE 0 0 0 > c0t3d0 ONLINE 0 0 0 > c0t7d0 ONLINE 0 0 0 > c0t11d0 ONLINE 0 0 0 > c0t15d0 ONLINE 0 0 0 > c0t19d0 ONLINE 0 0 0 > c0t23d0 ONLINE 0 0 0 > > errors: No known data errors > > > I tried doing a zpool clear with no luck: > > [EMAIL PROTECTED] ~]# zpool clear LogData c0t20d0 > [EMAIL PROTECTED] ~]# zpool status -v > pool: LogData > state: DEGRADED > status: One or more devices are faulted in response to persistent errors. > Sufficient replicas exist for the pool to continue functioning in a > degraded state. > action: Replace the faulted device, or use 'zpool clear' to mark the device > repaired. > scrub: resilver completed with 0 errors on Wed Feb 27 11:51:45 2008 > config: > > NAME STATE READ WRITE CKSUM > LogData DEGRADED 0 0 0 > raidz2 DEGRADED 0 0 0 > c0t12d0 ONLINE 0 0 0 > c0t5d0 ONLINE 0 0 0 > c0t0d0 ONLINE 0 0 0 > c0t4d0 ONLINE 0 0 0 > c0t8d0 ONLINE 0 0 0 > c0t16d0 ONLINE 0 0 0 > c0t20d0 ONLINE 0 0 0 > c0t1d0 ONLINE 0 0 0 > c0t9d0 ONLINE 0 0 0 > c0t13d0 ONLINE 0 0 0 > c0t17d0 ONLINE 0 0 0 > c0t20d0 FAULTED 0 0 0 too many errors > c0t2d0 ONLINE 0 0 0 > c0t6d0 ONLINE 0 0 0 > c0t10d0 ONLINE 0 0 0 > c0t14d0 ONLINE 0 0 0 > c0t18d0 ONLINE 0 0 0 > c0t22d0 ONLINE 0 0 0 > c0t3d0 ONLINE 0 0 0 > c0t7d0 ONLINE 0 0 0 > > And I’ve tried zpool replace: > > [EMAIL PROTECTED] ~]# > [EMAIL PROTECTED] ~]# zpool replace -f LogData c0t20d0 > invalid vdev specification > the following errors must be manually repaired: > /dev/dsk/c0t20d0s0 is part of active ZFS pool LogData. Please see zpool(1M). > > > So.. What am I missing here folks? > > Any help would be appreciated. >
Did you pull out the old drive and add a new one in its place hot? What does cfgadm -al report? Your drives should look like this: sata0/0::dsk/c7t0d0 disk connected configured ok sata0/1::dsk/c7t1d0 disk connected configured ok sata1/0::dsk/c8t0d0 disk connected configured ok sata1/1::dsk/c8t1d0 disk connected configured ok If c0t20d0 isn't configured, use # cfgadm -c configure sata1/1::dsk/c0t20d0 before attempting the zpool replace. hth - - Bart -- Bart Smaalders Solaris Kernel Performance [EMAIL PROTECTED] http://blogs.sun.com/barts "You will contribute more with mercurial than with thunderbird." _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss