Hi!

We also observe the same behavior on our test Hammer install, and I wrote about 
it some time ago:

<http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/22609>http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/22609

Jan Schremes give us some suggestions in thread, but we still not got any 
positive results - TCMalloc usage is
high. The usage is lowered to <10%, when disable crc in messages, disable debug 
and disable cephx auth,
but this is od course not for production use. Also we got a different trace, 
while performin FIO-RBD benchmarks
on ssd pool:
---
  46,07%  [kernel]              [k] _raw_spin_lock
   6,51%  [kernel]              [k] mb_cache_entry_alloc
   5,74%  libtcmalloc.so.4.2.2  [.] 
tcmalloc::CentralFreeList::FetchFromOneSpans(int, void**, void**)
   5,50%  libtcmalloc.so.4.2.2  [.] tcmalloc::SLL_Next(void*)
   3,86%  libtcmalloc.so.4.2.2  [.] TCMalloc_PageMap3<35>::get(unsigned long) 
const
   2,73%  libtcmalloc.so.4.2.2  [.] 
tcmalloc::CentralFreeList::ReleaseToSpans(void*)
   0,69%  libtcmalloc.so.4.2.2  [.] 
tcmalloc::CentralFreeList::ReleaseListToSpans(void*)
   0,69%  libtcmalloc.so.4.2.2  [.] tcmalloc::PageHeap::GetDescriptor(unsigned 
long) const
   0,64%  libtcmalloc.so.4.2.2  [.] tcmalloc::SLL_PopRange(void**, int, void**, 
void**)
---

I dont clearly understand, what's happening in this case: ssd pool is connected 
to the same host,
but different controller (C60X onboard instead of LSI2208), io scheduler set to 
noop, pool is gathered
from 4х400Gb Intel DC S3700 and have to perform better, I think - more than 
30-40 kops.
But we got the trace above and no more then 12-15 kiops. Where can be a problem?






Megov Igor
CIO, Yuterra

________________________________
От: ceph-users <ceph-users-boun...@lists.ceph.com> от имени YeYin <ey...@qq.com>
Отправлено: 17 августа 2015 г. 12:58
Кому: ceph-users
Тема: [ceph-users] tcmalloc use a lot of CPU

Hi, all,
  When I do performance test with rados bench, I found tcmalloc consumed a lot 
of CPU:

Samples: 265K of event 'cycles', Event count (approx.): 104385445900
+  27.58%  libtcmalloc.so.4.1.0    [.] 
tcmalloc::CentralFreeList::FetchFromSpans()
+  15.25%  libtcmalloc.so.4.1.0    [.] 
tcmalloc::ThreadCache::ReleaseToCentralCache(tcmalloc::ThreadCache::FreeList*, 
unsigned long,
+  12.20%  libtcmalloc.so.4.1.0    [.] 
tcmalloc::CentralFreeList::ReleaseToSpans(void*)
+   1.63%  perf                    [.] append_chain
+   1.39%  libtcmalloc.so.4.1.0    [.] 
tcmalloc::CentralFreeList::ReleaseListToSpans(void*)
+   1.02%  libtcmalloc.so.4.1.0    [.] 
tcmalloc::CentralFreeList::RemoveRange(void**, void**, int)
+   0.85%  libtcmalloc.so.4.1.0    [.] 0x0000000000017e6f
+   0.75%  libtcmalloc.so.4.1.0    [.] 
tcmalloc::ThreadCache::IncreaseCacheLimitLocked()
+   0.67%  libc-2.12.so            [.] memcpy
+   0.53%  libtcmalloc.so.4.1.0    [.] operator delete(void*)

Ceph version:
# ceph --version
ceph version 0.87.2 (87a7cec9ab11c677de2ab23a7668a77d2f5b955e)

Kernel version:
3.10.83

Is this phenomenon normal?Is there any idea about this problem?

Thanks.
Ye

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to