I'm in the process of replacing drive in a pool, and the resilver
times seem to have increased with each device. The way that I'm doing
this is by pulling a drive, physically replacing it, then doing
'cfgadm -c configure ____ ; zpool replace tank ____'. I don't have any
hot-swap bays available, so I'm physically replacing the device before
doing a 'zpool replace'.

I'm replacing Western Digital WD10EADS 1TB drives with Hitachi 5K3000
3TB drives. Neither device is fast, but they aren't THAT slow. wsvc_t
and asvc_t both look fairly healthy giving the device types.

Replacing the first device (took about 20 hours) went about as
expected. The second took about 44 hours. The third is still running
and should finish in slightly over 48 hours.

I'm wondering if the following would help for the next drive:
# zpool offline tank c2t4d0
# cfgadm -c unconfigure sata3/4::dsk/c2t4d0

At this point pull the drive and put it into an external USB adapter.
Put the new drive in the hot-swap bay. The USB adapter shows up as
c4t0d0.

# zpool online tank c4t0d0

This should re-add it to the pool and resilver the last few
transactions that may have been missed, right?

Then I want to actually replace the drive in the zpool:
# cfgadm -c configure sata3/4
# zpool replace tank c4t0d0 c2t4d0

Will this work? Will the replace go faster, since it won't need to
resilver from the parity data?


$ zpool list tank
NAME   SIZE  ALLOC   FREE    CAP  DEDUP  HEALTH  ALTROOT
tank  7.25T  6.40T   867G    88%  1.11x  DEGRADED  -
$ zpool status -x
  pool: tank
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
        continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
 scan: resilver in progress since Sat Apr 23 17:03:13 2011
    5.91T scanned out of 6.40T at 38.0M/s, 3h42m to go
    752G resilvered, 92.43% done
config:

        NAME              STATE     READ WRITE CKSUM
        tank              DEGRADED     0     0     0
          raidz2-0        DEGRADED     0     0     0
            c2t0d0        ONLINE       0     0     0
            c2t1d0        ONLINE       0     0     0
            c2t2d0        ONLINE       0     0     0
            c2t3d0        ONLINE       0     0     0
            c2t4d0        ONLINE       0     0     0
            replacing-5   DEGRADED     0     0     0
              c2t5d0/old  FAULTED      0     0     0  corrupted data
              c2t5d0      ONLINE       0     0     0  (resilvering)
            c2t6d0        ONLINE       0     0     0
            c2t7d0        ONLINE       0     0     0

errors: No known data errors
$ zpool iostat -v tank 60 3
                     capacity     operations    bandwidth
pool              alloc   free   read  write   read  write
----------------  -----  -----  -----  -----  -----  -----
tank              6.40T   867G    566     25  32.2M   156K
  raidz2          6.40T   867G    566     25  32.2M   156K
    c2t0d0            -      -    362     11  5.56M  71.6K
    c2t1d0            -      -    365     11  5.56M  71.6K
    c2t2d0            -      -    363     11  5.56M  71.6K
    c2t3d0            -      -    363     11  5.56M  71.6K
    c2t4d0            -      -    361     11  5.54M  71.6K
    replacing         -      -      0    492  8.28K  4.79M
      c2t5d0/old      -      -    202      5  2.84M  36.7K
      c2t5d0          -      -      0    315  8.66K  4.78M
    c2t6d0            -      -    170    190  2.68M  2.69M
    c2t7d0            -      -    386     10  5.53M  71.6K
----------------  -----  -----  -----  -----  -----  -----

                     capacity     operations    bandwidth
pool              alloc   free   read  write   read  write
----------------  -----  -----  -----  -----  -----  -----
tank              6.40T   867G    612     14  8.43M  70.7K
  raidz2          6.40T   867G    612     14  8.43M  70.7K
    c2t0d0            -      -    411     11  1.51M  57.9K
    c2t1d0            -      -    414     11  1.50M  58.0K
    c2t2d0            -      -    385     11  1.51M  57.9K
    c2t3d0            -      -    412     11  1.50M  58.0K
    c2t4d0            -      -    412     11  1.45M  57.8K
    replacing         -      -      0    574    366   852K
      c2t5d0/old      -      -      0      0      0      0
      c2t5d0          -      -      0    324    366   852K
    c2t6d0            -      -    427     11  1.45M  57.8K
    c2t7d0            -      -    431     11  1.49M  57.9K
----------------  -----  -----  -----  -----  -----  -----

                     capacity     operations    bandwidth
pool              alloc   free   read  write   read  write
----------------  -----  -----  -----  -----  -----  -----
tank              6.40T   867G  1.02K     12  11.1M  69.4K
  raidz2          6.40T   867G  1.02K     12  11.1M  69.4K
    c2t0d0            -      -    772     10  1.99M  59.3K
    c2t1d0            -      -    771     10  1.99M  59.4K
    c2t2d0            -      -    743     10  2.02M  59.4K
    c2t3d0            -      -    771     11  2.01M  59.3K
    c2t4d0            -      -    767     10  1.94M  59.1K
    replacing         -      -      0  1.00K     17  1.48M
      c2t5d0/old      -      -      0      0      0      0
      c2t5d0          -      -      0    533     17  1.48M
    c2t6d0            -      -    791     10  1.98M  59.2K
    c2t7d0            -      -    796     10  1.99M  59.3K
----------------  -----  -----  -----  -----  -----  -----

$ iostat -xn 60 3
                    extended device statistics
    r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
  362.4   11.5 5693.9   71.6  0.7  0.7    2.0    2.0  14  30 c2t0d0
  365.3   11.5 5689.0   71.6  0.7  0.7    1.8    1.9  14  29 c2t1d0
  363.2   11.5 5693.2   71.6  0.7  0.7    1.9    2.0  14  30 c2t2d0
  364.0   11.5 5692.7   71.6  0.7  0.7    1.9    1.9  14  30 c2t3d0
  361.2   11.5 5672.8   71.6  0.7  0.7    1.9    1.9  14  30 c2t4d0
  202.4  163.1 2915.2 2475.3  0.3  1.1    0.8    2.9   7  26 c2t5d0
  170.4  190.4 2747.3 2757.6  0.5  1.3    1.5    3.6  11  31 c2t6d0
  386.4   11.2 5659.0   71.6  0.5  0.6    1.3    1.5  12  27 c2t7d0
   95.0    1.2   94.5   16.1  0.0  0.0    0.2    0.2   0   1 c0t0d0
    0.9    1.2    3.3   16.1  0.0  0.0    7.5    1.9   0   0 c0t1d0
                    extended device statistics
    r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
  514.1   13.0 1937.7   65.7  0.2  0.8    0.3    1.5   5  27 c2t0d0
  510.1   13.2 1943.1   65.7  0.2  0.8    0.5    1.6   6  29 c2t1d0
  513.3   13.2 1926.3   65.8  0.2  0.8    0.3    1.5   5  28 c2t2d0
  505.9   13.3 1936.7   65.8  0.2  0.9    0.3    1.8   5  30 c2t3d0
  513.8   12.8 1890.1   65.8  0.2  0.8    0.3    1.5   5  26 c2t4d0
    0.1  488.6    0.1 1216.5  0.0  2.2    0.0    4.6   0  33 c2t5d0
  533.3   12.7 1875.3   65.9  0.1  0.7    0.2    1.3   4  24 c2t6d0
  541.6   12.9 1923.2   65.8  0.1  0.7    0.2    1.2   3  23 c2t7d0
    0.0    2.0    0.0    9.4  0.0  0.0    1.0    0.2   0   0 c0t0d0
    0.0    2.0    0.0    9.4  0.0  0.0    1.0    0.2   0   0 c0t1d0
                    extended device statistics
    r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
  506.7    9.2 1906.9   50.2  0.6  0.2    1.2    0.5  20  23 c2t0d0
  509.8    9.3 1909.5   50.2  0.6  0.2    1.2    0.4  19  23 c2t1d0
  508.6    9.0 1900.4   50.2  0.7  0.3    1.4    0.5  21  25 c2t2d0
  506.8    9.4 1897.2   50.3  0.6  0.2    1.2    0.5  19  23 c2t3d0
  505.1    9.4 1852.4   50.4  0.6  0.2    1.2    0.5  19  23 c2t4d0
    0.0  487.6    0.0 1227.9  0.0  3.5    0.0    7.2   0  46 c2t5d0
  534.8    9.2 1855.6   50.2  0.6  0.2    1.0    0.4  18  22 c2t6d0
  540.5    9.3 1891.4   50.2  0.5  0.2    1.0    0.4  17  21 c2t7d0
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t0d0
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t1d0



-- 
Brandon High : bh...@freaks.com
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to