Karun,
You can check how much data each OSD has with "ceph osd df"
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS
1 hdd 1.84000 1.00000 1885G 769G 1115G 40.84 0.97 101
3 hdd 4.64000 1.00000 4679G 2613G 2065G 55.86 1.33 275
4 hdd 4.64000 1.00000 4674G 1914G 2759G 40.96 0.97 193
5 hdd 4.64000 1.00000 4668G 1434G 3234G 30.72 0.73 148
8 hdd 1.84000 1.00000 1874G 742G 1131G 39.61 0.94 74
0 hdd 4.64000 1.00000 4668G 2331G 2337G 49.94 1.19 268
2 hdd 1.84000 1.00000 4668G 868G 3800G 18.60 0.44 99
6 hdd 4.64000 1.00000 4668G 2580G 2087G 55.28 1.32 275
7 hdd 1.84000 1.00000 1874G 888G 985G 47.43 1.13 107
TOTAL 33661G 14144G 19516G 42.02
MIN/MAX VAR: 0.44/1.33 STDDEV: 11.27
The "%USE" column shows how much space is used on each OSD. You may
need to change the weight of some of the OSDs so the data balances out
correctly with "ceph osd crush reweight osd.N W".Change the N to the
number of OSD and W to the new weight.
As you can see from above even though the weight on my 4.6TB is the
same for all of them, they have different %USE. So I could lower the
weight of the OSDs with more data, and Ceph will balance the cluster.
I am not too sure why this happens.
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-March/008623.html
Cary
-Dynamic
On Tue, Dec 19, 2017 at 3:57 PM, Jean-Charles Lopez <[email protected]> wrote:
> Hi
>
> did you set quotas on these pools?
>
> See this page for explanation of most error messages:
> http://docs.ceph.com/docs/master/rados/operations/health-checks/#pool-near-full
>
> JC
>
> On Dec 19, 2017, at 01:48, Karun Josy <[email protected]> wrote:
>
> Hello,
>
> In one of our clusters, health is showing these warnings :
> ---------
> OSD_NEARFULL 1 nearfull osd(s)
> osd.22 is near full
> POOL_NEARFULL 3 pool(s) nearfull
> pool 'templates' is nearfull
> pool 'cvm' is nearfull
> pool 'ecpool' is nearfull
> ------------
>
> One osd is above 85% used, which I know caused the OSD_Nearfull flag.
> But what does pool(s) nearfull mean ?
> And how can I correct it ?
>
> ]$ ceph df
> GLOBAL:
> SIZE AVAIL RAW USED %RAW USED
> 31742G 11147G 20594G 64.88
> POOLS:
> NAME ID USED %USED MAX AVAIL OBJECTS
> templates 5 196G 23.28 645G 50202
> cvm 6 6528 0 1076G 770
> ecpool 7 10260G 83.56 2018G 3004031
>
>
>
> Karun
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com