-----Original message-----
> From:Alwin Antreich <[email protected]>
> Sent: Thursday 6th September 2018 16:27
> To: ceph-users <[email protected]>
> Cc: Menno Zonneveld <[email protected]>
> Subject: Re: [ceph-users] Rados performance inconsistencies, lower than
> expected performance
>
> Hi,
Hi!
> On Thu, Sep 06, 2018 at 03:52:21PM +0200, Menno Zonneveld wrote:
> > ah yes, 3x replicated with minimal 2.
> >
> >
> > my ceph.conf is pretty bare, just in case it might be relevant
> >
> > [global]
> > auth client required = cephx
> > auth cluster required = cephx
> > auth service required = cephx
> >
> > cluster network = 172.25.42.0/24
> >
> > fsid = f4971cca-e73c-46bc-bb05-4af61d419f6e
> >
> > keyring = /etc/pve/priv/$cluster.$name.keyring
> >
> > mon allow pool delete = true
> > mon osd allow primary affinity = true
> On our test cluster, we didn't set the primary affinity as all OSDs were
> SSDs of the same model. Did you do any settings other than this? How
> does your crush map look like?
I only used this option when testing with mixing HDD and SSD (1 replica on SSD
and 2 on HDD); right now affinity for all disks is 1.
The weight of one OSD in each server is lower because I have partitioned the
drive to be able to test with SSD journal for HDDs but this isn't active at the
moment.
If I understand correctly setting the weight like this should be fine and I
also tested with weight 1 for all OSD's and I still get the same performance
('slow' when empty, fast when full)
Current ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 3.71997 root ssd
-5 1.23999 host ceph01-test
2 ssd 0.36600 osd.2 up 1.00000 1.00000
3 ssd 0.43700 osd.3 up 1.00000 1.00000
6 ssd 0.43700 osd.6 up 1.00000 1.00000
-7 1.23999 host ceph02-test
4 ssd 0.36600 osd.4 up 1.00000 1.00000
5 ssd 0.43700 osd.5 up 1.00000 1.00000
7 ssd 0.43700 osd.7 up 1.00000 1.00000
-3 1.23999 host ceph03-test
0 ssd 0.36600 osd.0 up 1.00000 1.00000
1 ssd 0.43700 osd.1 up 1.00000 1.00000
8 ssd 0.43700 osd.8 up 1.00000 1.00000
My current crush map looks like this:
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54
# devices
device 0 osd.0 class ssd
device 1 osd.1 class ssd
device 2 osd.2 class ssd
device 3 osd.3 class ssd
device 4 osd.4 class ssd
device 5 osd.5 class ssd
device 6 osd.6 class ssd
device 7 osd.7 class ssd
device 8 osd.8 class ssd
# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 region
type 10 root
# buckets
host ceph03-test {
id -3 # do not change unnecessarily
id -4 class ssd # do not change unnecessarily
# weight 1.240
alg straw2
hash 0 # rjenkins1
item osd.1 weight 0.437
item osd.0 weight 0.366
item osd.8 weight 0.437
}
host ceph01-test {
id -5 # do not change unnecessarily
id -6 class ssd # do not change unnecessarily
# weight 1.240
alg straw2
hash 0 # rjenkins1
item osd.3 weight 0.437
item osd.2 weight 0.366
item osd.6 weight 0.437
}
host ceph02-test {
id -7 # do not change unnecessarily
id -8 class ssd # do not change unnecessarily
# weight 1.240
alg straw2
hash 0 # rjenkins1
item osd.5 weight 0.437
item osd.4 weight 0.366
item osd.7 weight 0.437
}
root ssd {
id -1 # do not change unnecessarily
id -2 class ssd # do not change unnecessarily
# weight 3.720
alg straw2
hash 0 # rjenkins1
item ceph03-test weight 1.240
item ceph01-test weight 1.240
item ceph02-test weight 1.240
}
# rules
rule ssd {
id 0
type replicated
min_size 1
max_size 10
step take ssd
step chooseleaf firstn 0 type host
step emit
}
# end crush map
> >
> > osd journal size = 5120
> > osd pool default min size = 2
> > osd pool default size = 3
> >
> >
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com