Dear John,
Thank for your respone.
My cluster have 4616 pgs and 13 pools show as below:

in which default.rgw.buckets.data is erasure pool with config:
+ pg_num=pgp_num=1792
+ size=12
+ erause-code-profile:
   directory=/usr/lib64/ceph/erasure-code
   k=9
   m=3
   plugin=isa
   ruleset-failure-domain=osd
   ruleset-root=default

And all of other pools as: rgw.root, default.rgw.control ... are
replication pools with pg_num=pgp_num=256 and size=3.
These configurations above can affect system resources?
Here is my memory info:
Thank,
​
​

2017-02-15 19:00 GMT+07:00 John Petrini <[email protected]>:

> You should subtract buffers and cache from the used memory to get a more
> accurate representation of how much memory is actually available to
> processes. In this case that puts you around 22G of used - or a better term
> might be unavailable memory. Buffers and cache can be reallocated when
> needed - it's just Linux taking advantage of memory under the theory why
> not use it if it's there? Memory is fast so Linux will take advantage of it.
>
> With 72 OSD's 22G of memory puts you below the 500MB/daemon that you've
> mentioned so I don't think you have anything to be concerned about.
>
> ___
>
> John Petrini
>
> The information transmitted is intended only for the person or entity to
> which it is addressed and may contain confidential and/or privileged
> material. Any review, retransmission,  dissemination or other use of, or
> taking of any action in reliance upon, this information by persons or
> entities other than the intended recipient is prohibited. If you received
> this in error, please contact the sender and delete the material from any
> computer.
>
> On Tue, Feb 14, 2017 at 11:24 PM, Khang Nguyễn Nhật <
> [email protected]> wrote:
>
>> Hi Sam,
>> Thank for your reply. I use BTRFS file system on OSDs.
>> Here is result of "*free -hw*":
>>
>>                    total         used          free        shared
>> buffers       cache     available
>> Mem:           125G         58G         31G        1.2M        3.7M
>>   36G         60G
>>
>> and "*ceph df*":
>>
>> GLOBAL:
>>     SIZE     AVAIL     RAW USED     %RAW USED
>>     523T      522T        1539G          0.29
>> POOLS:
>>     NAME                           ID     USED     %USED     MAX AVAIL
>>   OBJECTS
>>     ............
>>     default.rgw.buckets.data  92     597G      0.15              391T
>>        84392
>>     ............
>>
>> I was reviced this a few minutes ago.
>>
>> 2017-02-15 10:50 GMT+07:00 Sam Huracan <[email protected]>:
>>
>>> Hi Khang,
>>>
>>> What file system do you use in OSD node?
>>> XFS always use Memory for caching data before writing to disk.
>>>
>>> So, don't worry, it always holds memory in your system as much as
>>> possible.
>>>
>>>
>>>
>>> 2017-02-15 10:35 GMT+07:00 Khang Nguyễn Nhật <
>>> [email protected]>:
>>>
>>>> Hi all,
>>>> My ceph OSDs is running on Fedora-server24 with config are:
>>>> 128GB RAM DDR3, CPU Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz, 72 OSDs
>>>> (8TB per OSD). My cluster was use ceph object gateway with S3 API. Now, it
>>>> had contained 500GB data but it was used > 50GB RAM. I'm worry my OSD will
>>>> dead if i continue put file to it. I had read "OSDs do not require as
>>>> much RAM for regular operations (e.g., 500MB of RAM per daemon instance);
>>>> however, during recovery they need significantly more RAM (e.g., ~1GB per
>>>> 1TB of storage per daemon)." in Ceph Hardware Recommendations. Someone
>>>> can give me advice on this issue? Thank
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> [email protected]
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to