There's some metadata on Bluestore OSDs (the rocksdb database), it's
usually ~1% of your data.
The DB will start out at a size of around 1GB, so that's expected.


Paul

2018-06-04 15:55 GMT+02:00 Marc-Antoine Desrochers <
[email protected]>:

> Hi,
>
>
>
> Im not sure if it’s normal or not but each time I add a new osd with
> ceph-deploy osd create --data /dev/sdg ceph-n1.
>
> It add 1GB to my global data but I just format the drive so it’s supposed
> to be at 0 right ?
>
> So I have 6 osd in my ceph and it took 6gib.
>
>
>
> [root@ceph-n1 ~]# ceph -s
>
>   cluster:
>
>     id:     1d97aa70-2029-463a-b6fa-20e98f3e21fb
>
>     health: HEALTH_OK
>
>
>
>   services:
>
>     mon: 1 daemons, quorum ceph-n1
>
>     mgr: ceph-n1(active)
>
>     mds: cephfs-1/1/1 up  {0=ceph-n1=up:active}
>
>     osd: 6 osds: 6 up, 6 in
>
>
>
>   data:
>
>     pools:   2 pools, 600 pgs
>
>     objects: 341 objects, 63109 kB
>
>     usage:   6324 MB used, 2782 GB / 2788 GB avail
>
>     pgs:     600 active+clean
>
>
>
>
>
> So im kind of confused...
>
> Thanks for your help.
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to