I'm in the process of replacing drive in a pool, and the resilver times seem to have increased with each device. The way that I'm doing this is by pulling a drive, physically replacing it, then doing 'cfgadm -c configure ____ ; zpool replace tank ____'. I don't have any hot-swap bays available, so I'm physically replacing the device before doing a 'zpool replace'.
I'm replacing Western Digital WD10EADS 1TB drives with Hitachi 5K3000 3TB drives. Neither device is fast, but they aren't THAT slow. wsvc_t and asvc_t both look fairly healthy giving the device types. Replacing the first device (took about 20 hours) went about as expected. The second took about 44 hours. The third is still running and should finish in slightly over 48 hours. I'm wondering if the following would help for the next drive: # zpool offline tank c2t4d0 # cfgadm -c unconfigure sata3/4::dsk/c2t4d0 At this point pull the drive and put it into an external USB adapter. Put the new drive in the hot-swap bay. The USB adapter shows up as c4t0d0. # zpool online tank c4t0d0 This should re-add it to the pool and resilver the last few transactions that may have been missed, right? Then I want to actually replace the drive in the zpool: # cfgadm -c configure sata3/4 # zpool replace tank c4t0d0 c2t4d0 Will this work? Will the replace go faster, since it won't need to resilver from the parity data? $ zpool list tank NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT tank 7.25T 6.40T 867G 88% 1.11x DEGRADED - $ zpool status -x pool: tank state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Sat Apr 23 17:03:13 2011 5.91T scanned out of 6.40T at 38.0M/s, 3h42m to go 752G resilvered, 92.43% done config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 c2t0d0 ONLINE 0 0 0 c2t1d0 ONLINE 0 0 0 c2t2d0 ONLINE 0 0 0 c2t3d0 ONLINE 0 0 0 c2t4d0 ONLINE 0 0 0 replacing-5 DEGRADED 0 0 0 c2t5d0/old FAULTED 0 0 0 corrupted data c2t5d0 ONLINE 0 0 0 (resilvering) c2t6d0 ONLINE 0 0 0 c2t7d0 ONLINE 0 0 0 errors: No known data errors $ zpool iostat -v tank 60 3 capacity operations bandwidth pool alloc free read write read write ---------------- ----- ----- ----- ----- ----- ----- tank 6.40T 867G 566 25 32.2M 156K raidz2 6.40T 867G 566 25 32.2M 156K c2t0d0 - - 362 11 5.56M 71.6K c2t1d0 - - 365 11 5.56M 71.6K c2t2d0 - - 363 11 5.56M 71.6K c2t3d0 - - 363 11 5.56M 71.6K c2t4d0 - - 361 11 5.54M 71.6K replacing - - 0 492 8.28K 4.79M c2t5d0/old - - 202 5 2.84M 36.7K c2t5d0 - - 0 315 8.66K 4.78M c2t6d0 - - 170 190 2.68M 2.69M c2t7d0 - - 386 10 5.53M 71.6K ---------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ---------------- ----- ----- ----- ----- ----- ----- tank 6.40T 867G 612 14 8.43M 70.7K raidz2 6.40T 867G 612 14 8.43M 70.7K c2t0d0 - - 411 11 1.51M 57.9K c2t1d0 - - 414 11 1.50M 58.0K c2t2d0 - - 385 11 1.51M 57.9K c2t3d0 - - 412 11 1.50M 58.0K c2t4d0 - - 412 11 1.45M 57.8K replacing - - 0 574 366 852K c2t5d0/old - - 0 0 0 0 c2t5d0 - - 0 324 366 852K c2t6d0 - - 427 11 1.45M 57.8K c2t7d0 - - 431 11 1.49M 57.9K ---------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ---------------- ----- ----- ----- ----- ----- ----- tank 6.40T 867G 1.02K 12 11.1M 69.4K raidz2 6.40T 867G 1.02K 12 11.1M 69.4K c2t0d0 - - 772 10 1.99M 59.3K c2t1d0 - - 771 10 1.99M 59.4K c2t2d0 - - 743 10 2.02M 59.4K c2t3d0 - - 771 11 2.01M 59.3K c2t4d0 - - 767 10 1.94M 59.1K replacing - - 0 1.00K 17 1.48M c2t5d0/old - - 0 0 0 0 c2t5d0 - - 0 533 17 1.48M c2t6d0 - - 791 10 1.98M 59.2K c2t7d0 - - 796 10 1.99M 59.3K ---------------- ----- ----- ----- ----- ----- ----- $ iostat -xn 60 3 extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 362.4 11.5 5693.9 71.6 0.7 0.7 2.0 2.0 14 30 c2t0d0 365.3 11.5 5689.0 71.6 0.7 0.7 1.8 1.9 14 29 c2t1d0 363.2 11.5 5693.2 71.6 0.7 0.7 1.9 2.0 14 30 c2t2d0 364.0 11.5 5692.7 71.6 0.7 0.7 1.9 1.9 14 30 c2t3d0 361.2 11.5 5672.8 71.6 0.7 0.7 1.9 1.9 14 30 c2t4d0 202.4 163.1 2915.2 2475.3 0.3 1.1 0.8 2.9 7 26 c2t5d0 170.4 190.4 2747.3 2757.6 0.5 1.3 1.5 3.6 11 31 c2t6d0 386.4 11.2 5659.0 71.6 0.5 0.6 1.3 1.5 12 27 c2t7d0 95.0 1.2 94.5 16.1 0.0 0.0 0.2 0.2 0 1 c0t0d0 0.9 1.2 3.3 16.1 0.0 0.0 7.5 1.9 0 0 c0t1d0 extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 514.1 13.0 1937.7 65.7 0.2 0.8 0.3 1.5 5 27 c2t0d0 510.1 13.2 1943.1 65.7 0.2 0.8 0.5 1.6 6 29 c2t1d0 513.3 13.2 1926.3 65.8 0.2 0.8 0.3 1.5 5 28 c2t2d0 505.9 13.3 1936.7 65.8 0.2 0.9 0.3 1.8 5 30 c2t3d0 513.8 12.8 1890.1 65.8 0.2 0.8 0.3 1.5 5 26 c2t4d0 0.1 488.6 0.1 1216.5 0.0 2.2 0.0 4.6 0 33 c2t5d0 533.3 12.7 1875.3 65.9 0.1 0.7 0.2 1.3 4 24 c2t6d0 541.6 12.9 1923.2 65.8 0.1 0.7 0.2 1.2 3 23 c2t7d0 0.0 2.0 0.0 9.4 0.0 0.0 1.0 0.2 0 0 c0t0d0 0.0 2.0 0.0 9.4 0.0 0.0 1.0 0.2 0 0 c0t1d0 extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 506.7 9.2 1906.9 50.2 0.6 0.2 1.2 0.5 20 23 c2t0d0 509.8 9.3 1909.5 50.2 0.6 0.2 1.2 0.4 19 23 c2t1d0 508.6 9.0 1900.4 50.2 0.7 0.3 1.4 0.5 21 25 c2t2d0 506.8 9.4 1897.2 50.3 0.6 0.2 1.2 0.5 19 23 c2t3d0 505.1 9.4 1852.4 50.4 0.6 0.2 1.2 0.5 19 23 c2t4d0 0.0 487.6 0.0 1227.9 0.0 3.5 0.0 7.2 0 46 c2t5d0 534.8 9.2 1855.6 50.2 0.6 0.2 1.0 0.4 18 22 c2t6d0 540.5 9.3 1891.4 50.2 0.5 0.2 1.0 0.4 17 21 c2t7d0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c0t0d0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c0t1d0 -- Brandon High : bh...@freaks.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss