That's right. I didn't actually use Jewel for very long. I'm glad it worked
for you.

On Fri, May 11, 2018, 4:49 PM Webert de Souza Lima <webert.b...@gmail.com>
wrote:

> Thanks David.
> Although you mentioned this was introduced with Luminous, it's working
> with Jewel.
>
> ~# ceph osd pool stats
>
>                             Fri May 11 17:41:39 2018
>
> pool rbd id 5
>   client io 505 kB/s rd, 3801 kB/s wr, 46 op/s rd, 27 op/s wr
>
> pool rbd_cache id 6
>   client io 2538 kB/s rd, 3070 kB/s wr, 601 op/s rd, 758 op/s wr
>   cache tier io 12225 kB/s flush, 0 op/s promote, 3 PG(s) flushing
>
> pool cephfs_metadata id 7
>   client io 2233 kB/s rd, 2260 kB/s wr, 95 op/s rd, 587 op/s wr
>
> pool cephfs_data_ssd id 8
>   client io 1126 kB/s rd, 94897 B/s wr, 33 op/s rd, 42 op/s wr
>
> pool cephfs_data id 9
>   client io 0 B/s rd, 11203 kB/s wr, 12 op/s rd, 12 op/s wr
>
> pool cephfs_data_cache id 10
>   client io 4383 kB/s rd, 550 kB/s wr, 57 op/s rd, 39 op/s wr
>   cache tier io 7012 kB/s flush, 4399 kB/s evict, 11 op/s promote
>
>
> Regards,
>
> Webert Lima
> DevOps Engineer at MAV Tecnologia
> *Belo Horizonte - Brasil*
> *IRC NICK - WebertRLZ*
>
>
> On Fri, May 11, 2018 at 5:14 PM David Turner <drakonst...@gmail.com>
> wrote:
>
>> `ceph osd pool stats` with the option to specify the pool you are
>> interested in should get you the breakdown of IO per pool.  This was
>> introduced with luminous.
>>
>> On Fri, May 11, 2018 at 2:39 PM Webert de Souza Lima <
>> webert.b...@gmail.com> wrote:
>>
>>> I think ceph doesn't have IO metrics will filters by pool right? I see
>>> IO metrics from clients only:
>>>
>>> ceph_client_io_ops
>>> ceph_client_io_read_bytes
>>> ceph_client_io_read_ops
>>> ceph_client_io_write_bytes
>>> ceph_client_io_write_ops
>>>
>>> and pool "byte" metrics, but not "io":
>>>
>>> ceph_pool(write/read)_bytes(_total)
>>>
>>> Regards,
>>>
>>> Webert Lima
>>> DevOps Engineer at MAV Tecnologia
>>> *Belo Horizonte - Brasil*
>>> *IRC NICK - WebertRLZ*
>>>
>>> On Wed, May 9, 2018 at 2:23 PM Webert de Souza Lima <
>>> webert.b...@gmail.com> wrote:
>>>
>>>> Hey Jon!
>>>>
>>>> On Wed, May 9, 2018 at 12:11 PM, John Spray <jsp...@redhat.com> wrote:
>>>>
>>>>> It depends on the metadata intensity of your workload.  It might be
>>>>> quite interesting to gather some drive stats on how many IOPS are
>>>>> currently hitting your metadata pool over a week of normal activity.
>>>>>
>>>>
>>>> Any ceph built-in tool for this? maybe ceph daemonperf (altoght I'm not
>>>> sure what I should be looking at).
>>>> My current SSD disks have 2 partitions.
>>>>  - One is used for cephfs cache tier pool,
>>>>  - The other is used for both:  cephfs meta-data pool and cephfs
>>>> data-ssd (this is an additional cephfs data pool with only ssds with file
>>>> layout for a specific direcotory to use it)
>>>>
>>>> Because of this, iostat shows me peaks of 12k IOPS in the metadata
>>>> partition, but this could definitely be IO for the data-ssd pool.
>>>>
>>>>
>>>>> If you are doing large file workloads, and the metadata mostly fits in
>>>>> RAM, then the number of IOPS from the MDS can be very, very low.  On
>>>>> the other hand, if you're doing random metadata reads from a small
>>>>> file workload where the metadata does not fit in RAM, almost every
>>>>> client read could generate a read operation, and each MDS could easily
>>>>> generate thousands of ops per second.
>>>>>
>>>>
>>>> I have yet to measure it the right way but I'd assume my metadata fits
>>>> in RAM (a few 100s of MB only).
>>>>
>>>> This is an email hosting cluster with dozens of thousands of users so
>>>> there are a lot of random reads and writes, but not too many small files.
>>>> Email messages are concatenated together in files up to 4MB in size
>>>> (when a rotation happens).
>>>> Most user operations are dovecot's INDEX operations and I will keep
>>>> index directory in a SSD-dedicaded pool.
>>>>
>>>>
>>>>
>>>>> Isolating metadata OSDs is useful if the data OSDs are going to be
>>>>> completely saturated: metadata performance will be protected even if
>>>>> clients are hitting the data OSDs hard.
>>>>>
>>>>
>>>> This seems to be the case.
>>>>
>>>>
>>>>> If "heavy write" means completely saturating the cluster, then sharing
>>>>> the OSDs is risky.  If "heavy write" just means that there are more
>>>>> writes than reads, then it may be fine if the metadata workload is not
>>>>> heavy enough to make good use of SSDs.
>>>>>
>>>>
>>>> Saturarion will only happen in peak workloads, not often. By heavy
>>>> write I mean there are much more writes than reads, yes.
>>>> So I think I can start sharing the OSDs, if I think this is impacting
>>>> performance I can just change the ruleset and move metadata to a SSD-only
>>>> pool, right?
>>>>
>>>>
>>>>> The way I'd summarise this is: in the general case, dedicated SSDs are
>>>>> the safe way to go -- they're intrinsically better suited to metadata.
>>>>> However, in some quite common special cases, the overall number of
>>>>> metadata ops is so low that the device doesn't matter.
>>>>
>>>>
>>>>
>>>> Thank you very much John!
>>>> Webert Lima
>>>> DevOps Engineer at MAV Tecnologia
>>>> Belo Horizonte - Brasil
>>>> IRC NICK - WebertRLZ
>>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to