Turns out i didn't do reply-all.

On Tue, Feb 16, 2016 at 9:18 AM, John Hogenmiller <j...@hogenmiller.net>
wrote:

> > And again - is dual Xeon's power enough for 60-disk node and Erasure
> Code?
>
>
> This is something I've been attempting to determine as well. I'm not yet
> getting
> I'm testing with some white-label hardware, but essentially supermicro
> 2twinu's with a pair of E5-2609 Xeons and 64GB of memory.  (
> http://www.supermicro.com/products/system/2U/6028/SYS-6028TR-HTFR.cfm).
> This is attached to DAEs with 60 x 6TB drives, in JBOD.
>
> Conversely, Supermicro sells a 72-disk OSD node, which Redhat considers a
> supported "reference architecture" device. The processors in those nodes
> are E5-269 12-core, vs what I have which is quad-core.
> http://www.supermicro.com/solutions/storage_ceph.cfm  (SSG-6048R-OSD432).*
> I would highly recommend reflecting on the supermicro hardware and using
> that as your reference as well*. If you could get an eval unit, use that
> to compare with the hardware you're working with.
>
> I currently have mine setup with 7 nodes, 60 OSDs each, radosgw running
> one each node, and 5 ceph monitors. I plan to move the monitors to their
> own dedicated hardware, and in reading, I may only need 3 to manage the 420
> OSDs.   *I am currently just setup for replication instead of EC*, though
> I want to redo this cluster to use EC. *Also, I am still trying to work
> out how much of an impact placement groups have on performance, and I may
> have a performance-hampering amount.*.
>
> We test the system using locust speaking S3 to the radosgw. Transactions
> are distributed equally across all 7 nodes and we track the statistics. We
> started first emulating 1000 users and got over 4Gbps, but load average on
> all nodes was in the mid-100s, and after 15 minutes we started getting
> socket timeouts. We stopped the test, let load settle, and started back at
> 100 users.  We've been running this test about 5 days now.  Load average on
> all nodes floats between 40 and 70. The nodes with ceph-mon running on them
> do not appear to be taxed any more than the ones without. The radosgw
> itself seems to take up a decent amount of cpu (running civetweb, no ssl).
>  iowait is non existent, everything appears to be cpu bound.
>
> At 1000 users, we had 4.3Gbps of PUTs and 2.2Gbps of GETs. Did not capture
> the TPS on that short test.
> At 100 users, we're pushing 2Gbps  in PUTs and 1.24Gpbs in GETs. Averaging
> 115 TPS.
>
> All in all, the speeds are not bad for a single rack, but the CPU
> utilization is a big concern. We're currently using other (proprietary)
> object storage platforms on this hardware configuration. They have their
> own set of issues, but CPU utilization is typically not the problem, even
> at higher utilization.
>
>
>
> root@ljb01:/home/ceph/rain-cluster# ceph status
>     cluster 4ebe7995-6a33-42be-bd4d-20f51d02ae45
>      health HEALTH_OK
>      monmap e5: 5 mons at {hail02-r01-06=
> 172.29.4.153:6789/0,hail02-r01-08=172.29.4.155:6789/0,rain02-r01-01=172.29.4.148:6789/0,rain02-r01-03=172.29.4.150:6789/0,rain02-r01-04=172.29.4.151:6789/0
> }
>             election epoch 86, quorum 0,1,2,3,4
> rain02-r01-01,rain02-r01-03,rain02-r01-04,hail02-r01-06,hail02-r01-08
>      osdmap e2543: 423 osds: 419 up, 419 in
>             flags sortbitwise
>       pgmap v676131: 33848 pgs, 14 pools, 50834 GB data, 29660 kobjects
>             149 TB used, 2134 TB / 2284 TB avail
>                33848 active+clean
>   client io 129 MB/s rd, 182 MB/s wr, 1562 op/s
>
>
>
>  # ceph-osd + ceph-mon + radosgw
> top - 13:29:22 up 40 days, 22:05,  1 user,  load average: 47.76, 47.33,
> 47.08
> Tasks: 1001 total,   7 running, 994 sleeping,   0 stopped,   0 zombie
> %Cpu(s): 39.2 us, 44.7 sy,  0.0 ni,  9.9 id,  2.4 wa,  0.0 hi,  3.7 si,
>  0.0 st
> KiB Mem:  65873180 total, 64818176 used,  1055004 free,     9324 buffers
> KiB Swap:  8388604 total,  7801828 used,   586776 free. 17610868 cached Mem
>
>     PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+
> COMMAND
>
>  178129 ceph      20   0 3066452 618060   5440 S  54.6  0.9   2678:49
> ceph-osd
>  218049 ceph      20   0 6261880 179704   2872 S  33.4  0.3   1852:14
> radosgw
>  165529 ceph      20   0 2915332 579064   4308 S  19.7  0.9 530:12.65
> ceph-osd
>  185193 ceph      20   0 2932696 585724   4412 S  19.1  0.9 545:20.31
> ceph-osd
>   52334 ceph      20   0 3030300 618868   4328 S  15.8  0.9 543:53.64
> ceph-osd
>   23124 ceph      20   0 3037740 607088   4440 S  15.2  0.9 461:03.98
> ceph-osd
>  154031 ceph      20   0 2982344 525428   4044 S  14.9  0.8 587:17.62
> ceph-osd
>  191278 ceph      20   0 2835208 570100   4700 S  14.9  0.9 547:11.66
> ceph-osd
>
>  # ceph-osd + radosgw (no ceph-mon)
>
>  top - 13:31:22 up 40 days, 22:06,  1 user,  load average: 64.25, 59.76,
> 58.17
> Tasks: 1015 total,   4 running, 1011 sleeping,   0 stopped,   0 zombie
> %Cpu0  : 24.2 us, 48.5 sy,  0.0 ni, 10.9 id,  1.2 wa,  0.0 hi, 15.2 si,
>  0.0 st
> %Cpu1  : 30.8 us, 49.7 sy,  0.0 ni, 13.5 id,  1.8 wa,  0.0 hi,  4.2 si,
>  0.0 st
> %Cpu2  : 33.9 us, 49.5 sy,  0.0 ni, 10.5 id,  2.1 wa,  0.0 hi,  3.9 si,
>  0.0 st
> %Cpu3  : 31.3 us, 52.4 sy,  0.0 ni, 10.5 id,  2.7 wa,  0.0 hi,  3.0 si,
>  0.0 st
> %Cpu4  : 34.7 us, 41.3 sy,  0.0 ni, 20.4 id,  3.3 wa,  0.0 hi,  0.3 si,
>  0.0 st
> %Cpu5  : 38.3 us, 36.7 sy,  0.0 ni, 21.0 id,  3.7 wa,  0.0 hi,  0.3 si,
>  0.0 st
> %Cpu6  : 36.6 us, 37.5 sy,  0.0 ni, 19.8 id,  6.1 wa,  0.0 hi,  0.0 si,
>  0.0 st
> %Cpu7  : 38.0 us, 38.0 sy,  0.0 ni, 19.5 id,  4.3 wa,  0.0 hi,  0.3 si,
>  0.0 st
> KiB Mem:  65873180 total, 61946688 used,  3926492 free,     1260 buffers
> KiB Swap:  8388604 total,  7080048 used,  1308556 free.  7910772 cached Mem
>
>     PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+
> COMMAND
>
>  108861 ceph      20   0 6283364 279464   3024 S  27.6  0.4   1684:30
> radosgw
>     546 root      20   0       0      0      0 R  23.4  0.0 579:55.28
> kswapd0
>  184953 ceph      20   0 2971100 576784   4348 S  23.1  0.9   1265:58
> ceph-osd
>  178967 ceph      20   0 2970500 594756   6000 S  18.9  0.9 505:27.13
> ceph-osd
>  184105 ceph      20   0 3096276 627944   7096 S  18.0  1.0 581:12.28
> ceph-osd
>   56073 ceph      20   0 2888244 542024   4836 S  13.5  0.8 530:41.86
> ceph-osd
>   55083 ceph      20   0 2819060 518500   5052 S  13.2  0.8 513:21.96
> ceph-osd
>  175578 ceph      20   0 3083096 630712   5564 S  13.2  1.0 591:10.91
> ceph-osd
>  180725 ceph      20   0 2915828 553240   4836 S  12.9  0.8 518:22.47
> ceph-osd
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to