Re: [ceph-users] ceph journal double writes?

2013-06-26 Thread Gregory Farnum
On Wednesday, June 26, 2013, Oliver Fuckner wrote:

 Hi,
 I am fairly new to ceph and just built my first 4 systems.

 I use:
 Supermicro X9SCL-Board with E3-1240 (4*3.4GHz) CPU and 32GB RAM
 LSI 9211-4i SAS HBA with 24 SATA disks and 2 SSDs (Intel 3700, 100GB), all
 connected through a 6GBit-SAS expander
 CentOS 6.4 with Kernel 2.6.32-358.11.1, 64bit
 ceph 0.61.4
 Intel 10GigEthernet NICs are used to connect the nodes together
 xfs is used on journal and osds


 The SSDs are configured in a mdadm raid1 and used for journals.
 The SSDs can write 400MBytes/sec each, but the sum of all disks is exactly
 half of it, 200MBytes/sec.




 So there are 2 journal writes for every write to the osd?


No.


 Is this expected behaviour? Why?


No, but at a guess your expanders aren't behaving properly. Alternatively,
your SSDs don't handle twelve write streams so well -- that's quite a lot
of oversubscription.

I would test the write behavior of your disks independently of Ceph (but
simultaneously!) and see what happens.
-Greg

This can be seen with real load and rados bench write -t 64


 Details:


 SSDs:
 mdadm creation:
 mdadm --create /dev/md2 --run --raid-devices=2 --level=raid1 --name=ssd
 /dev/sdc /dev/sdd
 mkfs.xfs -f -i size=2048 /dev/md2

 mount options in /etc/fstab:
 /dev/md2  /data/journal xfs rw,noatime,discard  1 2

 OSDs:
 mkfs.xfs -i size=2048 /dev/$24-disks
 ceph-osd -i $ID --mkfs --mkkey --mkjournal --osd-data /data/osd.slotx
 --osd-journal /data/journal/slotx


 Journals are limited to 2GByte per osd via ceph.conf


 Thanks,
  Oliver
 __**_
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Software Engineer #42 @ http://inktank.com | http://ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph journal double writes?

2013-06-26 Thread Gregory Farnum
On Wed, Jun 26, 2013 at 11:57 AM, Oliver Fuckner o...@fuckner.net wrote:
 How do I debug expander behaviour? I know lsiutil, but is there something
 like iostat for sas lanes/phys? Talking about oversubscription: What I
 really try is 24 streams to one SSD-mirror. So I will probably need more
 ssds, okay...

Just by testing simultaneous writes to all your drives and seeing what
happens, basically.

  I would test the write behavior of your disks independently of Ceph (but
  simultaneously!) and see what happens.

 well dd to the ssds also shows 400MByte/sec with 4MByte blocks.

Ah, they're in a RAID-1 mirror with all 24 going to it? Yeah, 24
simultaneous streams is beyond what a lot of controllers and drives
can handle very well. Try running multiple streams to the drives and
see what their average, deviation, and aggregate bandwidth is. :) I
expect that's the issue.
-Greg
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com