Den tors 9 maj 2019 kl 15:46 skrev Feng Zhang <[email protected]>:

>
> For erasure pool, suppose I have 10 nodes, each has 10 6TB drives, so
> in total 100 drives. I make a 4+2 erasure pool, failure domain is
> host/node. Then if one drive failed, (assume the 6TB is fully used),
> what the maximum speed the recovering process can have? Also suppose
> the cluster network is 10GbE, each disk has maximum 200MB/s sequential
> throughput.
>


I think IOPS is what I experience makes the largest impact during recovery
of spinning
drives, not speed of sequential perf, so when recoverying you will see
progress
(on older clusters at least) in terms of number of misplaced/degraded
objects like:

misplaced    3454/34989583 objects (0.0123%)

and the first number (here is my guess) moves at the speed of the IOPS of
the drives
being repaired, so if your drive(s) from which you rebuild can do 100 IOPS,
then the above
scenario will take ~34 seconds, even if that sizes and raw speeds should
indicate something
else about how fast they could move 0.0123% of your stored data.

As soon as you get super fast SSDs and NVMEs, the limit moves somewhere
else since they
have crazy IOPS numbers, and hence will repair lots faster, but if you have
only spinning
drives, then "don't hold your breath" is good advice, since it will take
longer than 6 TB divided
by 200MB/s (8h20m) if you are unlucky and other drives can't help out in
the rebuild.

Fancy fast WAL/DB/Journals probably help a lot here, since they do affect
the "iops"
you experience from your spin-drive OSDs.

-- 
May the most significant bit of your life be positive.
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to