Hi guys,

First of all I would like to welcome everyone as it's my first post to that
group.

I have got some general and basic CEPH questions (for you probably). I
found that the official documentation is not really the best of the best in
this subject and a lots of things are actually pretty confusing - some of
them completely different in the real world.

I'm just starting with CEPH and everything is just brand new to me. I have
got some questions from my initial findings so far.

I have got a cluster containing 3 monitors and 9 OSDs (1TB each) on 3
different nodes (so 3 OSDs on each node).

I created a test pool which is at the moment destroyed however when I was
testing some things and rebooted one of the monitors to see what happens
everything was fine but had the message "HEALTH_WARN pool test has too few
pgs". I increased numer of pg and pgp to 300 for that pool and everything
just started resyncing. Fair enough. However before I did that I had
something like:

cluster 8b7e94df-3a7f-4d45-8a7c-d61c0ad25478
     health HEALTH_WARN pool test has too few pgs
     monmap e1: 3 mons at {ceph-test-mon-01=
172.17.12.11:6789/0,ceph-test-mon-02=172.17.12.12:6789/0,ceph-test-mon-03=172.17.12.13:6789/0},
election epoch 6, quorum 0,1,2
ceph-test-mon-01,ceph-test-mon-02,ceph-test-mon-03

     osdmap e109: 9 osds: 9 up, 9
in

      pgmap v1580: 200 pgs, 4 pools, 40912 MB data, 10232
objects

            121 GB used, 7283 GB / 7404 GB
avail

                 200 active+clean

So the interesting part is actually here:
121 GB used, 7283 GB / 7404 GB avail

# ceph df
GLOBAL:

    SIZE      AVAIL     RAW USED     %RAW
USED
    8329G     8209G     120G
1.45

POOLS:
    NAME         ID     USED       %USED     OBJECTS
    data         0      0          0         0
    metadata     1      0          0         0
    rbd          2      139        0         2
    test        3      40912M     0.48      10230

and after seting pg and pgp_num to 300 I had this:

cluster
8b7e94df-3a7f-4d45-8a7c-d61c0ad25478

     health
HEALTH_OK

     monmap e1: 3 mons at {ceph-test-mon-01=
172.17.12.11:6789/0,ceph-test-mon-02=172.17.12.12:6789/0,ceph-test-mon-03=172.17.12.13:6789/0},
election epoch 12, quorum 0,1,2
ceph-test-mon-01,ceph-test-mon-02,ceph-test-mon-03

     osdmap e643: 9 osds: 9 up, 9
in

      pgmap v3399: 492 pgs, 4 pools, 1497 GB data, 374
kobjects

            120 GB used, 8209 GB / 8329 GB
avail

                 492 active+clean

# ceph df
GLOBAL:
    SIZE      AVAIL     RAW USED     %RAW USED
    8329G     8209G     120G         1.45

POOLS:
    NAME         ID     USED      %USED     OBJECTS
    data         0      0         0         0
    metadata     1      0         0         0
    rbd          2      139       0         2
    test        3      1497G     17.98     383525

So in the first example I don't get these allocation numbers. If the raw
space 120GB used why the pool size is 40GB ?

Also in the second case when I increased number of pd/pgp that size grew up
to 1.4TB.

So where is that relation and why the pool size grew up so much ?

Also is there any easy way to find how the space is being distributed and
what contains what ? I found that I can log to the OSD boxes and df shows
me the utilization. However this can be pretty complicated if I will have
multiple boxes - let me say more than 10.

Thank you for your help.

Regards,
TH
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to