I have a number of "ZFS backup" servers (about 1PB split between four
machines).

Some of them have 16x 18TB drives, but a couple have a mix of 12TB and 14TB
drives (because that's what we had).

All are running Ubuntu 20.04 LTS (with a snapshot build) or 22.04 LTS
(bundled).

We're just doing our first replace (actually swapping in a 16TB drive for a
12TB because that's all we have).

root@hqs1:~# zpool version
zfs-2.1.99-784_gae07fc139
zfs-kmod-2.1.99-784_gae07fc139
root@hqs1:~# zpool status
  pool: hqs1p1
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
  scan: resilver in progress since Mon Dec  5 15:19:31 2022
177T scanned at 2.49G/s, 175T issued at 2.46G/s, 177T total
7.95T resilvered, 98.97% done, 00:12:39 to go
config:

NAME                                      STATE     READ WRITE CKSUM
hqs1p1                                    DEGRADED     0     0     0
 draid2:3d:16c:1s-0                      DEGRADED     0     0     0
   780530e1-d2e4-0040-aa8b-8c7bed75a14a  ONLINE       0     0     0
 (resilvering)
   9c4428e8-d16f-3849-97d9-22fc441750dc  ONLINE       0     0     0
 (resilvering)
   0e148b1d-69a3-3345-9478-343ecf6b855d  ONLINE       0     0     0
 (resilvering)
   98208ffe-4b31-564f-832d-5744c809f163  ONLINE       0     0     0
 (resilvering)
   3ac46b0a-9c46-e14f-8137-69227f3a890a  ONLINE       0     0     0
 (resilvering)
   44e8f62f-5d49-c345-9c89-ac82926d42b7  ONLINE       0     0     0
 (resilvering)
   968dbacd-1d85-0b40-a1fc-977a09ac5aaa  ONLINE       0     0     0
 (resilvering)
   e7ca2666-1067-f54c-b723-b464fb0a5fa3  ONLINE       0     0     0
 (resilvering)
   318ff075-8860-e84e-8063-f77775f57a2d  ONLINE       0     0     0
 (resilvering)
   replacing-9                           DEGRADED     0     0     0
     2888151727045752617                 UNAVAIL      0     0     0  was
/dev/disk/by-partuuid/85fa9347-8359-4942-a20d-da1f6016ea48
     sdd                                 ONLINE       0     0     0
 (resilvering)
   fd69f284-d05d-f145-9bdb-0da8a72bf311  ONLINE       0     0     0
 (resilvering)
   f40f997a-33a1-2a4e-bb8d-64223c441f0f  ONLINE       0     0     0
 (resilvering)
   dbc35ea9-95d1-bd40-b79e-90d8a37079a6  ONLINE       0     0     0
 (resilvering)
   ac62bf3e-517e-a444-ae4f-a784b81cd14c  ONLINE       0     0     0
 (resilvering)
   d211031c-54d4-2443-853c-7e5c075b28ab  ONLINE       0     0     0
 (resilvering)
   06ba16e5-05cf-9b45-a267-510bfe98ceb1  ONLINE       0     0     0
 (resilvering)
 draid2:3d:6c:1s-1                       ONLINE       0     0     0
   be297802-095c-7d43-9132-360627ba8ceb  ONLINE       0     0     0
   e849981c-7316-cb47-b926-61d444790518  ONLINE       0     0     0
   bbc6d66d-38e1-c448-9d00-10ba7adcd371  ONLINE       0     0     0
   9fb44c95-5ea6-2347-ae97-38de283f45bf  ONLINE       0     0     0
   b212cae5-5068-8740-b120-0618ad459c1f  ONLINE       0     0     0
   8c771f6b-7d48-e744-9e25-847230fd2fdd  ONLINE       0     0     0
spares
 draid2-0-0                              AVAIL
 draid2-1-0                              AVAIL

errors: No known data errors
root@hqs1:~#

Here's a snippet of zpool iostat -v 1 ...

                                            capacity     operations
bandwidth
pool                                      alloc   free   read  write   read
 write
----------------------------------------  -----  -----  -----  -----  -----
 -----
hqs1p1                                     178T  67.1T  5.68K    259   585M
  144M
  draid2:3d:16c:1s-0                       128T  63.1T  5.66K    259   585M
  144M
    780530e1-d2e4-0040-aa8b-8c7bed75a14a      -      -    208      0  25.4M
     0
    9c4428e8-d16f-3849-97d9-22fc441750dc      -      -   1010      2   102M
 23.7K
    0e148b1d-69a3-3345-9478-343ecf6b855d      -      -    145      0  34.1M
 15.8K
    98208ffe-4b31-564f-832d-5744c809f163      -      -    101      1  28.3M
 7.90K
    3ac46b0a-9c46-e14f-8137-69227f3a890a      -      -    511      0  53.5M
     0
    44e8f62f-5d49-c345-9c89-ac82926d42b7      -      -     12      0  4.82M
     0
    968dbacd-1d85-0b40-a1fc-977a09ac5aaa      -      -     22      0  5.43M
 15.8K
    e7ca2666-1067-f54c-b723-b464fb0a5fa3      -      -    227      2  36.7M
 23.7K
    318ff075-8860-e84e-8063-f77775f57a2d      -      -    999      1  83.1M
 7.90K
    replacing-9                               -      -      0    243      0
  144M
      2888151727045752617                     -      -      0      0      0
     0
      sdd                                     -      -      0    243      0
  144M
    fd69f284-d05d-f145-9bdb-0da8a72bf311      -      -    306      0  54.5M
 15.8K
    f40f997a-33a1-2a4e-bb8d-64223c441f0f      -      -     47      0  16.9M
     0
    dbc35ea9-95d1-bd40-b79e-90d8a37079a6      -      -    234      0  16.7M
 15.8K
    ac62bf3e-517e-a444-ae4f-a784b81cd14c      -      -    417      0  15.9M
     0
    d211031c-54d4-2443-853c-7e5c075b28ab      -      -    911      0  48.4M
 15.8K
    06ba16e5-05cf-9b45-a267-510bfe98ceb1      -      -    643      0  60.1M
 15.8K
  draid2:3d:6c:1s-1                       50.6T  3.96T     19      0   198K
     0
    be297802-095c-7d43-9132-360627ba8ceb      -      -      3      0  39.5K
     0
    e849981c-7316-cb47-b926-61d444790518      -      -      2      0  23.7K
     0
    bbc6d66d-38e1-c448-9d00-10ba7adcd371      -      -      2      0  23.7K
     0
    9fb44c95-5ea6-2347-ae97-38de283f45bf      -      -      3      0  39.5K
     0
    b212cae5-5068-8740-b120-0618ad459c1f      -      -      3      0  39.5K
     0
    8c771f6b-7d48-e744-9e25-847230fd2fdd      -      -      1      0  31.6K
     0
----------------------------------------  -----  -----  -----  -----  -----
 -----

Lots of DRAID goodness there. Seems to be resilvering a q good whack.

My main question is: why have the DRAID vdevs been so disproportionately
allocated?

------------------------------------------
openzfs: openzfs-developer
Permalink: 
https://openzfs.topicbox.com/groups/developer/T386dac0e170785f1-Ma99e75ef1c6e8ee3074003a0
Delivery options: https://openzfs.topicbox.com/groups/developer/subscription

Reply via email to