Re: [ceph-users] Unexpected data

2018-06-04 Thread Paul Emmerich
There's some metadata on Bluestore OSDs (the rocksdb database), it's
usually ~1% of your data.
The DB will start out at a size of around 1GB, so that's expected.


Paul

2018-06-04 15:55 GMT+02:00 Marc-Antoine Desrochers <
marc-antoine.desroch...@sogetel.com>:

> Hi,
>
>
>
> Im not sure if it’s normal or not but each time I add a new osd with
> ceph-deploy osd create --data /dev/sdg ceph-n1.
>
> It add 1GB to my global data but I just format the drive so it’s supposed
> to be at 0 right ?
>
> So I have 6 osd in my ceph and it took 6gib.
>
>
>
> [root@ceph-n1 ~]# ceph -s
>
>   cluster:
>
> id: 1d97aa70-2029-463a-b6fa-20e98f3e21fb
>
> health: HEALTH_OK
>
>
>
>   services:
>
> mon: 1 daemons, quorum ceph-n1
>
> mgr: ceph-n1(active)
>
> mds: cephfs-1/1/1 up  {0=ceph-n1=up:active}
>
> osd: 6 osds: 6 up, 6 in
>
>
>
>   data:
>
> pools:   2 pools, 600 pgs
>
> objects: 341 objects, 63109 kB
>
> usage:   6324 MB used, 2782 GB / 2788 GB avail
>
> pgs: 600 active+clean
>
>
>
>
>
> So im kind of confused...
>
> Thanks for your help.
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Unexpected data

2018-06-04 Thread Marc-Antoine Desrochers
Hi,

 

Im not sure if it's normal or not but each time I add a new osd with
ceph-deploy osd create --data /dev/sdg ceph-n1.

It add 1GB to my global data but I just format the drive so it's supposed to
be at 0 right ?

So I have 6 osd in my ceph and it took 6gib.

 

[root@ceph-n1 ~]# ceph -s

  cluster:

id: 1d97aa70-2029-463a-b6fa-20e98f3e21fb

health: HEALTH_OK

 

  services:

mon: 1 daemons, quorum ceph-n1

mgr: ceph-n1(active)

mds: cephfs-1/1/1 up  {0=ceph-n1=up:active}

osd: 6 osds: 6 up, 6 in

 

  data:

pools:   2 pools, 600 pgs

objects: 341 objects, 63109 kB

usage:   6324 MB used, 2782 GB / 2788 GB avail

pgs: 600 active+clean

 

 

So im kind of confused...

Thanks for your help.

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com