> Op 9 februari 2017 om 9:13 schreef Özhan Rüzgar Karaman
> <[email protected]>:
>
>
> Hi;
> I am using Hammer 0.49.9 release on my Ceph Storage, today i noticed that
> listing an rbd pool takes to much time then the old days. If i have more
> rbd images on pool it takes much more time.
>
It is the -l flag that you are using in addition. That flag opens each RBD
image and stats the header of it to get the size.
A regular 'rbd ls' will only read the RADOS object rbd_directory, but it is the
-l flag which causes the RBD tool to iterate over all the images and query
their header.
> My clusters health is ok and currently there is no load on the cluster.
> Only rbd images are used to serve to vm's.
>
> I am sending some information below. My level.db size is also 280 mb, i
> also compacted level.db to 40 mb size but again "rbd ls -l" output is too
> slow.
>
> This timing is important for my vm deploy time to complete because when i
> refresh a pool/datastore it takes nearly to 20 seconds or more for 350 rbd
> images+snapshots.
>
> Thanks for all help
>
> Regards
> Ozhan Ruzgar
>
> root@mont3:/var/lib/ceph/mon/ceph-mont3/store.db# ceph -s
> cluster 6b1cb3f4-85e6-4b70-b057-ba7716f823cc
> health HEALTH_OK
> monmap e1: 3 mons at
> {mont1=172.16.x.x:6789/0,mont2=172.16.x.x:6789/0,mont3=172.16.x.x:6789/0}
> election epoch 126, quorum 0,1,2 mont1,mont2,mont3
> osdmap e20509: 40 osds: 40 up, 40 in
> pgmap v20333442: 1536 pgs, 3 pools, 235 GB data, 63442 objects
> 700 GB used, 3297 GB / 3998 GB avail
> 1536 active+clean
> client io 0 B/s rd, 3785 kB/s wr, 314 op/s
>
> root@mont1:~# time rbd ls -l cst2|wc -l
> 278
>
> real 0m11.970s
> user 0m0.572s
> sys 0m0.316s
> root@mont1:~# time rbd ls -l cst3|wc -l
> 15
>
> real 0m0.396s
> user 0m0.020s
> sys 0m0.032s
> root@mont1:~# time rbd ls -l cst4|wc -l
> 330
>
> real 0m16.630s
> user 0m0.668s
> sys 0m0.336s
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com