Dne 9.5.2014 9:08, Christian Balzer napsal(a):
Is that really just one disk?

Yes, its just one disk in all PCs. I know that the setup is bad, but I
want just to get
familiar with Ceph (and other parallel fs like Gluster ot Lustre) and
see what they can
do and cannot.

You have the reason for the write performance half right.
Every write goes to the primary OSD of the PG for that object.
That is, the journal of that OSD, which in your configuration I suspect is
a file on the same XFS as the actual OSD data. Either way, it would be on
the same disk as you only have one.
So that write goes to the primary OSD journal, then gets replicated to the
journal of the secondary OSD, then it get's ACK'ed to the client.
Meanwhile the journals will have to get written to the actual storage
eventually.

So the client PC writes on one OSD to the journal and then the OSD
replicates the data
from that journal to the second OSD also into its journal. Only after
that the data on each OSD are
copied from the journals into actual OSD storage ? Interesting in that
case klient should write
around 100MB/s to one OSD then it should stop and wait for that OSD to
replicate dato onto
second OSD (also around 100MB/s) and then all is done. After the disk
with journlas and storage
space should copy all the data on itselfs.

So the journal is some kind of cache for OSDs?

From the graphs i got it seems that the klient is sending data to both
OSD in parallel into the journals.
Then each of the OSDs copy the data once more on itselfs (not sure). But
i dont know why the network
traffic has these spikes. Is it because the client writes some chunk of
data and then waits for something
before next chunk can be sent ?

So each write happens basically twice, your single disk now only has an
effective speed of around 60-70MB/s (couldn't find any benchmarks for your
model, but most drives of this type have write speeds up to 140MB/s).

They can work around 100MB/s for sure.

Now add to this the fact that the replication from the other OSD will of
course also impact things.
That network bandwidth for the replication has to come from somewhere...

Look at what the recommended configurations by Inktank are and at previous
threads in here to get an idea what helps.

Since I doubt you have the budget or parts for your test setup to add more
disks, SSDs for journals, HW cache controllers, additional network cards
and so forth I guess you will have to live with this performance for now.

Christian

-- Christian Balzer Network/Systems Engineer [email protected] Global
OnLine Japan/Fusion Communications http://www.gol.com/

I will do that. Thank you very much for your reply!


---
Tato zpráva neobsahuje viry ani jiný škodlivý kód - avast! Antivirus je aktivní.
http://www.avast.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to