Hi Christian,
How are you? Thanks a lot for the answers, mine in red.
--- Original message ---
Asunto: Re: [ceph-users] Using Ramdisk wi
De: Christian Balzer <[email protected]>
Para: <[email protected]>
Cc: German Anders <[email protected]>
Fecha: Wednesday, 30/07/2014 11:42
Hello,
On Wed, 30 Jul 2014 09:55:49 -0400 German Anders wrote:
Hi Wido,
How are you? Thanks a lot for the quick response. I know
that is
heavy cost on using ramdisk, but also i want to try that to see if i
could get better performance, since I'm using a 10GbE network with the
following configuration and i can't achieve more than 300MB/s of
throughput on rbd:
Testing the limits of Ceph with a ramdisk based journal to see what is
possible in terms of speed (and you will find that it is CPU/protocol
bound) is fine.
Anything resembling production is a big no-no.
Got it, did you try flashcache from facebook or dm-cache?
MON Servers (3):
2x Intel Xeon E3-1270v3 @3.5Ghz (8C)
32GB RAM
2x SSD Intel 120G in RAID1 for OS
1x 10GbE port
OSD Servers (4):
2x Intel Xeon E5-2609v2 @2.5Ghz (8C)
64GB RAM
2x SSD Intel 120G in RAID1 for OS
3x SSD Intel 120G for Journals (3 SAS disks: 1 SSD
Journal)
You're not telling us WHICH actual Intel SSDs you're using.
If those are DC3500 ones, then 300MB/s totoal isn't a big surprise at
all,
as they are capable of 135MB/s writes at most.
The SSD model is Intel SSDSC2BB120G4 firm D2010370
9x SAS 3TB 6G for OSD
That would be somewhere over 1GB/s in theory, but give file system and
other overheads (what is your replication level?) that's a very
theoretical value indeed.
The RF is 2, so perf should be much better, also notice that read perf
is really poor, around 62MB/s...
Christian
2x 10GbE port (1 for Cluster Network, 1 for Public
Network)
- 10GbE Switches (1 for Cluster interconnect and 1 for Public network)
- Using Ceph Firefly version 0.80.4.
The thing is that with fio, rados bench and vdbench tools
we
only see 300MB/s on writes (rand and seq) with bs of 4m and 16
threads, that's pretty low actually, yesterday i was talking in the
ceph irc and i hit with the presentation that someone from Fujitsu do
on Frankfurt and also with some mails with some config at 10GbE and
he achieve almost 795MB/s and more... i would like to know if possible
how to implement that so we could improve our ceph cluster a little
bit more, i actually configure the scheduler on the SSD's disks both
OS and Journal to [noop] but still didn't notice any improvement.
That's why we would like to try RAMDISK on Journals, i've noticed that
he implement that on their Ceph cluster.
I will really appreciate the help on this. Also if you need me to send
you some more information about the Ceph scheme please let me know.
Also if someone could share some detail conf info will really help!
Thanks a lot,
German Anders
--- Original message ---
Asunto: Re: [ceph-users] Using Ramdisk wi
De: Wido den Hollander <[email protected]>
Para: <[email protected]>
Fecha: Wednesday, 30/07/2014 10:34
On 07/30/2014 03:28 PM, German Anders wrote:
Hi Everyone,
Anybody is using ramdisk to put the
Journal on it? If
so, could
you please share the commands to implement that? since I'm having some
issues with that and want to test that out to see if i could get
better
performance.
Don't do this. When you loose the journal, you loose the OSD. So a
reboot of the machine effectively trashes the data on that OSD.
Wido
Thanks in advance,
*German Anders
*
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
Wido den Hollander
42on B.V.
Ceph trainer and consultant
Phone: +31 (0)20 700 9902
Skype: contact42on
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
Christian Balzer Network/Systems Engineer
[email protected] Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com