Hi all,

I have a test ceph cluster with 4 osd servers containing each 3 osds.

The crushmap uses 2 rooms with 2 servers in each room.

We use replica 3 for pools.

I have the following custom crushrule to ensure that I have at least one copy of each data in each room.

rule replicated3over2rooms {
    id 1
    type replicated
    min_size 3
    max_size 4
    step take default
    step choose firstn 0 type room
    step chooseleaf firstn 2 type host
    step emit
}

Everything was working well in nautilus/centos7 (I could create pools using the dashboard and my custom rule).

I upgraded to pacific/ubuntu 20.04 in containers with cephadm.

Now, I cannot create a new pool with replicated3over2rooms using the dashboard !

If I choose Pool type = replicated, Replicated size = 3, Crush ruleset = replicated3over2rooms

The dashboard says :

Minimum: 3
Maximum: 2
The size specified is out of range. A value from 3 to 2 is usable.

And inspecting  replicatedover2rooms ruleset in the dashboard says that the parameters are

max_size 4
min_size 3
rule_id 1
usable_size 2

Where that usable_size comes from ? How to correct it ?

If i run the command line

ceph osd pool create test 16 replicated   replicated3over2rooms 3

it works !!

Thanks.

F.



_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to