Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-04-01 Thread German Anders

Hi everyone,

 I want to know the scheme if someone has a Ceph Cluster and 
running on it Cassandra databases, and if possible to send some 
performance numbers. Also if i wanna know if someone has a linux 
server (Ubuntu or CentOS) running on an HP Blade server and use a Ceph 
cluster as Backend with 10GbE ports?, if that's the case, if it's 
possible to run this DD and send me the output to see how it performs:


dd if=/dev/zero of=./$RANDOM bs=4k count=22 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=8k count=14 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=16k count=9 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=32k count=4 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=64k count=2 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=128k count=1 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=256k count=4000 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=512k count=3000 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=1024k count=3000 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=2048k count=1000 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=4096k count=800 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=8192k count=400 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=16384k count=200 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=32768k count=300 oflag=direct
dd if=/dev/zero of=./$RANDOM bs=65536k count=260 oflag=direct

And also if possible to run a sysbench:

sysbench --num-threads=16 --test=fileio --file-total-size=3G 
--file-test-mode=rndrw prepare
sysbench --num-threads=16 --test=fileio --file-total-size=3G 
--file-test-mode=rndrw run
sysbench --num-threads=16 --test=fileio --file-total-size=3G 
--file-test-mode=rndrw cleanup


Thanks in advance,

Best regards,



German Anders











--- Original message ---
Asunto: Re: [ceph-users] RBD as backend for iSCSI SAN Targets
De: Jianing Yang 
Para: Karol Kozubal 
Cc: ceph-users@lists.ceph.com 
Fecha: Tuesday, 01/04/2014 07:39


On Fri 28 Mar 2014 08:55:30 AM CST, Karol Kozubal wrote:


Hi Jianing,



Sorry for the late reply, I missed your contribution to the thread.


> Thank you for your response. I am still waiting for some of my 
hardware
> and will begin testing the new setup with firefly once it is 
available as
> a long term support release. I am looking forward to testing the new 
setup.


> I am curious about more details on your proxy node configuration for 
the
> tgt deamons? I am interested if your setup tolerates node failure on 
the

iscsi end of things, if so how it is configured?



Actually, The fault tolerance is provided by ms exchange servers. We
just setup two proxy nodes (server /w tgt daemon). One for master
database and the other for backup database. The exchange servers will 
do

the switch thing on failure.


Thanks,



Karol






> On 2014-03-19, 6:58 AM, "Jianing Yang"  
wrote:



>Hi, Karol
>
> >Here is something that I can share. We are running Ceph as an 
Exchange
> >Backend via iSCSI. We currently host about 2000 mailboxes which is 
about

>7 TB data overall. Our configuration is
>
>- Proxy Node (with tgt daemon) x 2
>- Ceph Monitor x 3 (virtual machines)
> >- Ceph OSD x 50 (SATA 7200rpm 2T), Replica = 2, Journal on OSD (I 
know it

>is
>bad, but ...)
>
> >We tested RBD using fio and got a randwrite around 1500 iops. On 
the

>living system, I saw the highest op/s around 3.1k.
>
> >I've benchmarked "tgt with librdb" vs "tgt with kernel rbd" using 
my

>virtual machines. It seems that "tgt with librdb" doesn't perform
>well. It has only 1/5 iops of kernel rbd.
>
> >We are new to Ceph and still finding ways to improve the 
performance. I

>am really looking forward to your benchmark.
>
>On Sun 16 Mar 2014 12:40:53 AM CST, Karol Kozubal wrote:
>
> > Hi Wido,
>
> > > I will have some new hardware for running tests in the next two 
weeks

>or
> > > so and will report my findings once I get a chance to run some 
tests. I
> > > will disable writeback on the target side as I will be 
attempting to
> > > configure an ssd caching pool of 24 ssd's with writeback for the 
main

>pool
> > > with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I 
will be

> > running everything through 10Gig SFP+ Ethernet interfaces with a
>dedicated
> > cluster network interface, dedicated public ceph interface and a
>separate
> > > iscsi network also with 10 gig interfaces for the target 
machines.

>
> > > I am ideally looking for a 20,000 to 60,000 IOPS from this 
system if I

>can
> > > get the caching pool configuration right. The application has a 
30ms

>max
> > latency requirement for the storage.
>
> > > In my current tests I have only spinners with SAS 10K disks, 
4.2ms

>write
> > > latency on the disks with separ

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-04-01 Thread Jianing Yang

On Fri 28 Mar 2014 08:55:30 AM CST, Karol Kozubal wrote:

 > Hi Jianing,

 > Sorry for the late reply, I missed your contribution to the thread.

 > Thank you for your response. I am still waiting for some of my hardware
 > and will begin testing the new setup with firefly once it is available as
 > a long term support release. I am looking forward to testing the new setup.

 > I am curious about more details on your proxy node configuration for the
 > tgt deamons? I am interested if your setup tolerates node failure on the
 > iscsi end of things, if so how it is configured?


Actually, The fault tolerance is provided by ms exchange servers. We
just setup two proxy nodes (server /w tgt daemon). One for master
database and the other for backup database. The exchange servers will do
the switch thing on failure.

 > Thanks,

 > Karol





 > On 2014-03-19, 6:58 AM, "Jianing Yang"  wrote:

 > >Hi, Karol
 > >
 > >Here is something that I can share. We are running Ceph as an Exchange
 > >Backend via iSCSI. We currently host about 2000 mailboxes which is about
 > >7 TB data overall. Our configuration is
 > >
 > >- Proxy Node (with tgt daemon) x 2
 > >- Ceph Monitor x 3 (virtual machines)
 > >- Ceph OSD x 50 (SATA 7200rpm 2T), Replica = 2, Journal on OSD (I know it
 > >is
 > >bad, but ...)
 > >
 > >We tested RBD using fio and got a randwrite around 1500 iops. On the
 > >living system, I saw the highest op/s around 3.1k.
 > >
 > >I've benchmarked "tgt with librdb" vs "tgt with kernel rbd" using my
 > >virtual machines. It seems that "tgt with librdb" doesn't perform
 > >well. It has only 1/5 iops of kernel rbd.
 > >
 > >We are new to Ceph and still finding ways to improve the performance. I
 > >am really looking forward to your benchmark.
 > >
 > >On Sun 16 Mar 2014 12:40:53 AM CST, Karol Kozubal wrote:
 > >
 > > > Hi Wido,
 > >
 > > > I will have some new hardware for running tests in the next two weeks
 > >or
 > > > so and will report my findings once I get a chance to run some tests. I
 > > > will disable writeback on the target side as I will be attempting to
 > > > configure an ssd caching pool of 24 ssd's with writeback for the main
 > >pool
 > > > with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
 > > > running everything through 10Gig SFP+ Ethernet interfaces with a
 > >dedicated
 > > > cluster network interface, dedicated public ceph interface and a
 > >separate
 > > > iscsi network also with 10 gig interfaces for the target machines.
 > >
 > > > I am ideally looking for a 20,000 to 60,000 IOPS from this system if I
 > >can
 > > > get the caching pool configuration right. The application has a 30ms
 > >max
 > > > latency requirement for the storage.
 > >
 > > > In my current tests I have only spinners with SAS 10K disks, 4.2ms
 > >write
 > > > latency on the disks with separate journaling on SAS 15K disks with a
 > > > 3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned
 > >with
 > > > the overall operation apply latency that I have been seeing (1-6ms
 > >idle is
 > > > normal, but up to 60-170ms for a moderate workload using rbd
 > >bench-write)
 > > > however I am on a network where I am bound to 1500 mtu and I will get
 > >to
 > > > test jumbo frames with the next setup in addition to the ssd¹s. I
 > >suspect
 > > > the overall performance will be good in the new test setup and I am
 > > > curious to see what my tests will yield.
 > >
 > > > Thanks for the response!
 > >
 > > > Karol
 > >
 > >
 > >
 > > > On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:
 > >
 > > > >On 03/15/2014 04:11 PM, Karol Kozubal wrote:
 > > > >> Hi Everyone,
 > > > >>
 > > > >> I am just wondering if any of you are running a ceph cluster with an
 > > > >> iSCSI target front end? I know this isn¹t available out of the box,
 > > > >> unfortunately in one particular use case we are looking at providing
 > > > >> iSCSI access and it's a necessity. I am liking the idea of having
 > >rbd
 > > > >> devices serving block level storage to the iSCSI Target servers
 > >while
 > > > >> providing a unified backed for native rbd access by openstack and
 > > > >> various application servers. On multiple levels this would reduce
 > >the
 > > > >> complexity of our SAN environment and move us away from expensive
 > > > >> proprietary solutions that don¹t scale out.
 > > > >>
 > > > >> If any of you have deployed any HA iSCSI Targets backed by rbd I
 > >would
 > > > >> really appreciate your feedback and any thoughts.
 > > > >>
 > > > >
 > > > >I haven't used it in production, but a couple of things which come to
 > > > >mind:
 > > > >
 > > > >- Use TGT so you can run it all in userspace backed by librbd
 > > > >- Do not use writeback caching on the targets
 > > > >
 > > > >You could use multipathing if you don't use writeback caching. Use
 > > > >writeback would also cause data loss/corruption in case of multiple
 > > > >targets.
 > > > >
 > > > >It will probably just work with TGT, but I don't know an

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-27 Thread Karol Kozubal
Hi Jianing,

Sorry for the late reply, I missed your contribution to the thread.

Thank you for your response. I am still waiting for some of my hardware
and will begin testing the new setup with firefly once it is available as
a long term support release. I am looking forward to testing the new setup.

I am curious about more details on your proxy node configuration for the
tgt deamons? I am interested if your setup tolerates node failure on the
iscsi end of things, if so how it is configured?

Thanks,

Karol





On 2014-03-19, 6:58 AM, "Jianing Yang"  wrote:

>Hi, Karol
>
>Here is something that I can share. We are running Ceph as an Exchange
>Backend via iSCSI. We currently host about 2000 mailboxes which is about
>7 TB data overall. Our configuration is
>
>- Proxy Node (with tgt daemon) x 2
>- Ceph Monitor x 3 (virtual machines)
>- Ceph OSD x 50 (SATA 7200rpm 2T), Replica = 2, Journal on OSD (I know it
>is
>bad, but ...)
>
>We tested RBD using fio and got a randwrite around 1500 iops. On the
>living system, I saw the highest op/s around 3.1k.
>
>I've benchmarked "tgt with librdb" vs "tgt with kernel rbd" using my
>virtual machines. It seems that "tgt with librdb" doesn't perform
>well. It has only 1/5 iops of kernel rbd.
>
>We are new to Ceph and still finding ways to improve the performance. I
>am really looking forward to your benchmark.
>
>On Sun 16 Mar 2014 12:40:53 AM CST, Karol Kozubal wrote:
>
> > Hi Wido,
>
> > I will have some new hardware for running tests in the next two weeks
>or
> > so and will report my findings once I get a chance to run some tests. I
> > will disable writeback on the target side as I will be attempting to
> > configure an ssd caching pool of 24 ssd's with writeback for the main
>pool
> > with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
> > running everything through 10Gig SFP+ Ethernet interfaces with a
>dedicated
> > cluster network interface, dedicated public ceph interface and a
>separate
> > iscsi network also with 10 gig interfaces for the target machines.
>
> > I am ideally looking for a 20,000 to 60,000 IOPS from this system if I
>can
> > get the caching pool configuration right. The application has a 30ms
>max
> > latency requirement for the storage.
>
> > In my current tests I have only spinners with SAS 10K disks, 4.2ms
>write
> > latency on the disks with separate journaling on SAS 15K disks with a
> > 3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned
>with
> > the overall operation apply latency that I have been seeing (1-6ms
>idle is
> > normal, but up to 60-170ms for a moderate workload using rbd
>bench-write)
> > however I am on a network where I am bound to 1500 mtu and I will get
>to
> > test jumbo frames with the next setup in addition to the ssd¹s. I
>suspect
> > the overall performance will be good in the new test setup and I am
> > curious to see what my tests will yield.
>
> > Thanks for the response!
>
> > Karol
>
>
>
> > On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:
>
> > >On 03/15/2014 04:11 PM, Karol Kozubal wrote:
> > >> Hi Everyone,
> > >>
> > >> I am just wondering if any of you are running a ceph cluster with an
> > >> iSCSI target front end? I know this isn¹t available out of the box,
> > >> unfortunately in one particular use case we are looking at providing
> > >> iSCSI access and it's a necessity. I am liking the idea of having
>rbd
> > >> devices serving block level storage to the iSCSI Target servers
>while
> > >> providing a unified backed for native rbd access by openstack and
> > >> various application servers. On multiple levels this would reduce
>the
> > >> complexity of our SAN environment and move us away from expensive
> > >> proprietary solutions that don¹t scale out.
> > >>
> > >> If any of you have deployed any HA iSCSI Targets backed by rbd I
>would
> > >> really appreciate your feedback and any thoughts.
> > >>
> > >
> > >I haven't used it in production, but a couple of things which come to
> > >mind:
> > >
> > >- Use TGT so you can run it all in userspace backed by librbd
> > >- Do not use writeback caching on the targets
> > >
> > >You could use multipathing if you don't use writeback caching. Use
> > >writeback would also cause data loss/corruption in case of multiple
> > >targets.
> > >
> > >It will probably just work with TGT, but I don't know anything about
>the
> > >performance.
> > >
> > >> Karol
> > >>
> > >>
> > >> ___
> > >> ceph-users mailing list
> > >> ceph-users@lists.ceph.com
> > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > >>
> > >
> > >
> > >--
> > >Wido den Hollander
> > >42on B.V.
> > >
> > >Phone: +31 (0)20 700 9902
> > >Skype: contact42on
> > >___
> > >ceph-users mailing list
> > >ceph-users@lists.ceph.com
> > >http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> > ___
> > ceph-users mailing list
> 

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-19 Thread Jianing Yang
Hi, Karol

Here is something that I can share. We are running Ceph as an Exchange
Backend via iSCSI. We currently host about 2000 mailboxes which is about
7 TB data overall. Our configuration is

- Proxy Node (with tgt daemon) x 2
- Ceph Monitor x 3 (virtual machines)
- Ceph OSD x 50 (SATA 7200rpm 2T), Replica = 2, Journal on OSD (I know it is
bad, but ...)

We tested RBD using fio and got a randwrite around 1500 iops. On the
living system, I saw the highest op/s around 3.1k.

I've benchmarked "tgt with librdb" vs "tgt with kernel rbd" using my
virtual machines. It seems that "tgt with librdb" doesn't perform
well. It has only 1/5 iops of kernel rbd.

We are new to Ceph and still finding ways to improve the performance. I
am really looking forward to your benchmark.

On Sun 16 Mar 2014 12:40:53 AM CST, Karol Kozubal wrote:

 > Hi Wido,

 > I will have some new hardware for running tests in the next two weeks or
 > so and will report my findings once I get a chance to run some tests. I
 > will disable writeback on the target side as I will be attempting to
 > configure an ssd caching pool of 24 ssd's with writeback for the main pool
 > with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
 > running everything through 10Gig SFP+ Ethernet interfaces with a dedicated
 > cluster network interface, dedicated public ceph interface and a separate
 > iscsi network also with 10 gig interfaces for the target machines.

 > I am ideally looking for a 20,000 to 60,000 IOPS from this system if I can
 > get the caching pool configuration right. The application has a 30ms max
 > latency requirement for the storage.

 > In my current tests I have only spinners with SAS 10K disks, 4.2ms write
 > latency on the disks with separate journaling on SAS 15K disks with a
 > 3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned with
 > the overall operation apply latency that I have been seeing (1-6ms idle is
 > normal, but up to 60-170ms for a moderate workload using rbd bench-write)
 > however I am on a network where I am bound to 1500 mtu and I will get to
 > test jumbo frames with the next setup in addition to the ssd¹s. I suspect
 > the overall performance will be good in the new test setup and I am
 > curious to see what my tests will yield.

 > Thanks for the response!

 > Karol



 > On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:

 > >On 03/15/2014 04:11 PM, Karol Kozubal wrote:
 > >> Hi Everyone,
 > >>
 > >> I am just wondering if any of you are running a ceph cluster with an
 > >> iSCSI target front end? I know this isn¹t available out of the box,
 > >> unfortunately in one particular use case we are looking at providing
 > >> iSCSI access and it's a necessity. I am liking the idea of having rbd
 > >> devices serving block level storage to the iSCSI Target servers while
 > >> providing a unified backed for native rbd access by openstack and
 > >> various application servers. On multiple levels this would reduce the
 > >> complexity of our SAN environment and move us away from expensive
 > >> proprietary solutions that don¹t scale out.
 > >>
 > >> If any of you have deployed any HA iSCSI Targets backed by rbd I would
 > >> really appreciate your feedback and any thoughts.
 > >>
 > >
 > >I haven't used it in production, but a couple of things which come to
 > >mind:
 > >
 > >- Use TGT so you can run it all in userspace backed by librbd
 > >- Do not use writeback caching on the targets
 > >
 > >You could use multipathing if you don't use writeback caching. Use
 > >writeback would also cause data loss/corruption in case of multiple
 > >targets.
 > >
 > >It will probably just work with TGT, but I don't know anything about the
 > >performance.
 > >
 > >> Karol
 > >>
 > >>
 > >> ___
 > >> ceph-users mailing list
 > >> ceph-users@lists.ceph.com
 > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
 > >>
 > >
 > >
 > >--
 > >Wido den Hollander
 > >42on B.V.
 > >
 > >Phone: +31 (0)20 700 9902
 > >Skype: contact42on
 > >___
 > >ceph-users mailing list
 > >ceph-users@lists.ceph.com
 > >http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 > ___
 > ceph-users mailing list
 > ceph-users@lists.ceph.com
 > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


--
 _
/ Save time starting to type a command or \
| file name, then press tab to complete   |
| Hit tab twice to bring up multiple  |
\ completion options. /
 -
   \
\
 _   ___
  ]OO|_n_n__][.  | |
   []_|__|)< | |
oooo  'oo -| oo\\_   ~~~|~~~
+--+--+--+--+--+--+--+--+--+--+--+--+--+
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-15 Thread Sage Weil
On Sat, 15 Mar 2014, Karol Kozubal wrote:
> I just re-read the documentation… It looks like its a proposed feature
> that is in development. I will have to adjust my test in consequence in
> that case.
> 
> Any one out there have any ideas when this will be implemented? Or what
> the plans look like as of right now?

This will appear in 0.78, which will be out in the next week.

sage

> 
> 
> 
> On 2014-03-15, 1:17 PM, "Karol Kozubal"  wrote:
> 
> >How are the SSDs going to be in writeback? Is that the new caching pool
> >Feature?
> >
> >I am not sure what version implemented this, but it is documented here
> >(https://ceph.com/docs/master/dev/cache-pool/).
> >I will be using the latest stable release for my next batch of testing,
> >right now I am on 0.67.4 and I will be moving towards the 0.72.x branch.
> >
> >As for the IOPS, it would be a total cluster IO throughput estimate based
> >on an application that would be reading/writing to more than 60 rbd
> >volumes.
> >
> >
> >
> >
> >
> >On 2014-03-15, 1:11 PM, "Wido den Hollander"  wrote:
> >
> >>On 03/15/2014 05:40 PM, Karol Kozubal wrote:
> >>> Hi Wido,
> >>>
> >>> I will have some new hardware for running tests in the next two weeks
> >>>or
> >>> so and will report my findings once I get a chance to run some tests. I
> >>> will disable writeback on the target side as I will be attempting to
> >>> configure an ssd caching pool of 24 ssd's with writeback for the main
> >>>pool
> >>> with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
> >>
> >>How are the SSDs going to be in writeback? Is that the new caching pool
> >>feature?
> >>
> >>> running everything through 10Gig SFP+ Ethernet interfaces with a
> >>>dedicated
> >>> cluster network interface, dedicated public ceph interface and a
> >>>separate
> >>> iscsi network also with 10 gig interfaces for the target machines.
> >>>
> >>
> >>That seems like a good network.
> >>
> >>> I am ideally looking for a 20,000 to 60,000 IOPS from this system if I
> >>>can
> >>> get the caching pool configuration right. The application has a 30ms
> >>>max
> >>> latency requirement for the storage.
> >>>
> >>
> >>20.000 to 60.000 is a big difference. But the only way you are going to
> >>achieve that is by doing a lot of parellel I/O. Ceph doesn't excel in
> >>single threads doing a lot of I/O.
> >>
> >>So if you have multiple RBD devices on which you are doing the I/O it
> >>shouldn't be that much of a problem.
> >>
> >>Just spread out the I/O. Scale horizontal instead of vertical.
> >>
> >>> In my current tests I have only spinners with SAS 10K disks, 4.2ms
> >>>write
> >>> latency on the disks with separate journaling on SAS 15K disks with a
> >>> 3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned
> >>>with
> >>> the overall operation apply latency that I have been seeing (1-6ms idle
> >>>is
> >>> normal, but up to 60-170ms for a moderate workload using rbd
> >>>bench-write)
> >>> however I am on a network where I am bound to 1500 mtu and I will get
> >>>to
> >>> test jumbo frames with the next setup in addition to the ssd¹s. I
> >>>suspect
> >>> the overall performance will be good in the new test setup and I am
> >>> curious to see what my tests will yield.
> >>>
> >>> Thanks for the response!
> >>>
> >>> Karol
> >>>
> >>>
> >>>
> >>> On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:
> >>>
>  On 03/15/2014 04:11 PM, Karol Kozubal wrote:
> > Hi Everyone,
> >
> > I am just wondering if any of you are running a ceph cluster with an
> > iSCSI target front end? I know this isn¹t available out of the box,
> > unfortunately in one particular use case we are looking at providing
> > iSCSI access and it's a necessity. I am liking the idea of having rbd
> > devices serving block level storage to the iSCSI Target servers while
> > providing a unified backed for native rbd access by openstack and
> > various application servers. On multiple levels this would reduce the
> > complexity of our SAN environment and move us away from expensive
> > proprietary solutions that don¹t scale out.
> >
> > If any of you have deployed any HA iSCSI Targets backed by rbd I
> >would
> > really appreciate your feedback and any thoughts.
> >
> 
>  I haven't used it in production, but a couple of things which come to
>  mind:
> 
>  - Use TGT so you can run it all in userspace backed by librbd
>  - Do not use writeback caching on the targets
> 
>  You could use multipathing if you don't use writeback caching. Use
>  writeback would also cause data loss/corruption in case of multiple
>  targets.
> 
>  It will probably just work with TGT, but I don't know anything about
> the
>  performance.
> 
> > Karol
> >
> >
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-15 Thread Karol Kozubal
I just re-read the documentation… It looks like its a proposed feature
that is in development. I will have to adjust my test in consequence in
that case.

Any one out there have any ideas when this will be implemented? Or what
the plans look like as of right now?



On 2014-03-15, 1:17 PM, "Karol Kozubal"  wrote:

>How are the SSDs going to be in writeback? Is that the new caching pool
>Feature?
>
>I am not sure what version implemented this, but it is documented here
>(https://ceph.com/docs/master/dev/cache-pool/).
>I will be using the latest stable release for my next batch of testing,
>right now I am on 0.67.4 and I will be moving towards the 0.72.x branch.
>
>As for the IOPS, it would be a total cluster IO throughput estimate based
>on an application that would be reading/writing to more than 60 rbd
>volumes.
>
>
>
>
>
>On 2014-03-15, 1:11 PM, "Wido den Hollander"  wrote:
>
>>On 03/15/2014 05:40 PM, Karol Kozubal wrote:
>>> Hi Wido,
>>>
>>> I will have some new hardware for running tests in the next two weeks
>>>or
>>> so and will report my findings once I get a chance to run some tests. I
>>> will disable writeback on the target side as I will be attempting to
>>> configure an ssd caching pool of 24 ssd's with writeback for the main
>>>pool
>>> with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
>>
>>How are the SSDs going to be in writeback? Is that the new caching pool
>>feature?
>>
>>> running everything through 10Gig SFP+ Ethernet interfaces with a
>>>dedicated
>>> cluster network interface, dedicated public ceph interface and a
>>>separate
>>> iscsi network also with 10 gig interfaces for the target machines.
>>>
>>
>>That seems like a good network.
>>
>>> I am ideally looking for a 20,000 to 60,000 IOPS from this system if I
>>>can
>>> get the caching pool configuration right. The application has a 30ms
>>>max
>>> latency requirement for the storage.
>>>
>>
>>20.000 to 60.000 is a big difference. But the only way you are going to
>>achieve that is by doing a lot of parellel I/O. Ceph doesn't excel in
>>single threads doing a lot of I/O.
>>
>>So if you have multiple RBD devices on which you are doing the I/O it
>>shouldn't be that much of a problem.
>>
>>Just spread out the I/O. Scale horizontal instead of vertical.
>>
>>> In my current tests I have only spinners with SAS 10K disks, 4.2ms
>>>write
>>> latency on the disks with separate journaling on SAS 15K disks with a
>>> 3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned
>>>with
>>> the overall operation apply latency that I have been seeing (1-6ms idle
>>>is
>>> normal, but up to 60-170ms for a moderate workload using rbd
>>>bench-write)
>>> however I am on a network where I am bound to 1500 mtu and I will get
>>>to
>>> test jumbo frames with the next setup in addition to the ssd¹s. I
>>>suspect
>>> the overall performance will be good in the new test setup and I am
>>> curious to see what my tests will yield.
>>>
>>> Thanks for the response!
>>>
>>> Karol
>>>
>>>
>>>
>>> On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:
>>>
 On 03/15/2014 04:11 PM, Karol Kozubal wrote:
> Hi Everyone,
>
> I am just wondering if any of you are running a ceph cluster with an
> iSCSI target front end? I know this isn¹t available out of the box,
> unfortunately in one particular use case we are looking at providing
> iSCSI access and it's a necessity. I am liking the idea of having rbd
> devices serving block level storage to the iSCSI Target servers while
> providing a unified backed for native rbd access by openstack and
> various application servers. On multiple levels this would reduce the
> complexity of our SAN environment and move us away from expensive
> proprietary solutions that don¹t scale out.
>
> If any of you have deployed any HA iSCSI Targets backed by rbd I
>would
> really appreciate your feedback and any thoughts.
>

 I haven't used it in production, but a couple of things which come to
 mind:

 - Use TGT so you can run it all in userspace backed by librbd
 - Do not use writeback caching on the targets

 You could use multipathing if you don't use writeback caching. Use
 writeback would also cause data loss/corruption in case of multiple
 targets.

 It will probably just work with TGT, but I don't know anything about
the
 performance.

> Karol
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


 --
 Wido den Hollander
 42on B.V.

 Phone: +31 (0)20 700 9902
 Skype: contact42on
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>>
>>-- 
>>Wido den Hollander
>>42on B.V.
>>
>>Phon

Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-15 Thread Karol Kozubal
How are the SSDs going to be in writeback? Is that the new caching pool
Feature?

I am not sure what version implemented this, but it is documented here
(https://ceph.com/docs/master/dev/cache-pool/).
I will be using the latest stable release for my next batch of testing,
right now I am on 0.67.4 and I will be moving towards the 0.72.x branch.

As for the IOPS, it would be a total cluster IO throughput estimate based
on an application that would be reading/writing to more than 60 rbd
volumes.





On 2014-03-15, 1:11 PM, "Wido den Hollander"  wrote:

>On 03/15/2014 05:40 PM, Karol Kozubal wrote:
>> Hi Wido,
>>
>> I will have some new hardware for running tests in the next two weeks or
>> so and will report my findings once I get a chance to run some tests. I
>> will disable writeback on the target side as I will be attempting to
>> configure an ssd caching pool of 24 ssd's with writeback for the main
>>pool
>> with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
>
>How are the SSDs going to be in writeback? Is that the new caching pool
>feature?
>
>> running everything through 10Gig SFP+ Ethernet interfaces with a
>>dedicated
>> cluster network interface, dedicated public ceph interface and a
>>separate
>> iscsi network also with 10 gig interfaces for the target machines.
>>
>
>That seems like a good network.
>
>> I am ideally looking for a 20,000 to 60,000 IOPS from this system if I
>>can
>> get the caching pool configuration right. The application has a 30ms max
>> latency requirement for the storage.
>>
>
>20.000 to 60.000 is a big difference. But the only way you are going to
>achieve that is by doing a lot of parellel I/O. Ceph doesn't excel in
>single threads doing a lot of I/O.
>
>So if you have multiple RBD devices on which you are doing the I/O it
>shouldn't be that much of a problem.
>
>Just spread out the I/O. Scale horizontal instead of vertical.
>
>> In my current tests I have only spinners with SAS 10K disks, 4.2ms write
>> latency on the disks with separate journaling on SAS 15K disks with a
>> 3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned
>>with
>> the overall operation apply latency that I have been seeing (1-6ms idle
>>is
>> normal, but up to 60-170ms for a moderate workload using rbd
>>bench-write)
>> however I am on a network where I am bound to 1500 mtu and I will get to
>> test jumbo frames with the next setup in addition to the ssd¹s. I
>>suspect
>> the overall performance will be good in the new test setup and I am
>> curious to see what my tests will yield.
>>
>> Thanks for the response!
>>
>> Karol
>>
>>
>>
>> On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:
>>
>>> On 03/15/2014 04:11 PM, Karol Kozubal wrote:
 Hi Everyone,

 I am just wondering if any of you are running a ceph cluster with an
 iSCSI target front end? I know this isn¹t available out of the box,
 unfortunately in one particular use case we are looking at providing
 iSCSI access and it's a necessity. I am liking the idea of having rbd
 devices serving block level storage to the iSCSI Target servers while
 providing a unified backed for native rbd access by openstack and
 various application servers. On multiple levels this would reduce the
 complexity of our SAN environment and move us away from expensive
 proprietary solutions that don¹t scale out.

 If any of you have deployed any HA iSCSI Targets backed by rbd I would
 really appreciate your feedback and any thoughts.

>>>
>>> I haven't used it in production, but a couple of things which come to
>>> mind:
>>>
>>> - Use TGT so you can run it all in userspace backed by librbd
>>> - Do not use writeback caching on the targets
>>>
>>> You could use multipathing if you don't use writeback caching. Use
>>> writeback would also cause data loss/corruption in case of multiple
>>> targets.
>>>
>>> It will probably just work with TGT, but I don't know anything about
>>>the
>>> performance.
>>>
 Karol


 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

>>>
>>>
>>> --
>>> Wido den Hollander
>>> 42on B.V.
>>>
>>> Phone: +31 (0)20 700 9902
>>> Skype: contact42on
>>> ___
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>-- 
>Wido den Hollander
>42on B.V.
>
>Phone: +31 (0)20 700 9902
>Skype: contact42on
>___
>ceph-users mailing list
>ceph-users@lists.ceph.com
>http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-15 Thread Wido den Hollander

On 03/15/2014 05:40 PM, Karol Kozubal wrote:

Hi Wido,

I will have some new hardware for running tests in the next two weeks or
so and will report my findings once I get a chance to run some tests. I
will disable writeback on the target side as I will be attempting to
configure an ssd caching pool of 24 ssd's with writeback for the main pool
with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be


How are the SSDs going to be in writeback? Is that the new caching pool 
feature?



running everything through 10Gig SFP+ Ethernet interfaces with a dedicated
cluster network interface, dedicated public ceph interface and a separate
iscsi network also with 10 gig interfaces for the target machines.



That seems like a good network.


I am ideally looking for a 20,000 to 60,000 IOPS from this system if I can
get the caching pool configuration right. The application has a 30ms max
latency requirement for the storage.



20.000 to 60.000 is a big difference. But the only way you are going to 
achieve that is by doing a lot of parellel I/O. Ceph doesn't excel in 
single threads doing a lot of I/O.


So if you have multiple RBD devices on which you are doing the I/O it 
shouldn't be that much of a problem.


Just spread out the I/O. Scale horizontal instead of vertical.


In my current tests I have only spinners with SAS 10K disks, 4.2ms write
latency on the disks with separate journaling on SAS 15K disks with a
3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned with
the overall operation apply latency that I have been seeing (1-6ms idle is
normal, but up to 60-170ms for a moderate workload using rbd bench-write)
however I am on a network where I am bound to 1500 mtu and I will get to
test jumbo frames with the next setup in addition to the ssd¹s. I suspect
the overall performance will be good in the new test setup and I am
curious to see what my tests will yield.

Thanks for the response!

Karol



On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:


On 03/15/2014 04:11 PM, Karol Kozubal wrote:

Hi Everyone,

I am just wondering if any of you are running a ceph cluster with an
iSCSI target front end? I know this isn¹t available out of the box,
unfortunately in one particular use case we are looking at providing
iSCSI access and it's a necessity. I am liking the idea of having rbd
devices serving block level storage to the iSCSI Target servers while
providing a unified backed for native rbd access by openstack and
various application servers. On multiple levels this would reduce the
complexity of our SAN environment and move us away from expensive
proprietary solutions that don¹t scale out.

If any of you have deployed any HA iSCSI Targets backed by rbd I would
really appreciate your feedback and any thoughts.



I haven't used it in production, but a couple of things which come to
mind:

- Use TGT so you can run it all in userspace backed by librbd
- Do not use writeback caching on the targets

You could use multipathing if you don't use writeback caching. Use
writeback would also cause data loss/corruption in case of multiple
targets.

It will probably just work with TGT, but I don't know anything about the
performance.


Karol


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




--
Wido den Hollander
42on B.V.

Phone: +31 (0)20 700 9902
Skype: contact42on
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





--
Wido den Hollander
42on B.V.

Phone: +31 (0)20 700 9902
Skype: contact42on
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-15 Thread Karol Kozubal
Hi Wido,

I will have some new hardware for running tests in the next two weeks or
so and will report my findings once I get a chance to run some tests. I
will disable writeback on the target side as I will be attempting to
configure an ssd caching pool of 24 ssd's with writeback for the main pool
with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
running everything through 10Gig SFP+ Ethernet interfaces with a dedicated
cluster network interface, dedicated public ceph interface and a separate
iscsi network also with 10 gig interfaces for the target machines.

I am ideally looking for a 20,000 to 60,000 IOPS from this system if I can
get the caching pool configuration right. The application has a 30ms max
latency requirement for the storage.

In my current tests I have only spinners with SAS 10K disks, 4.2ms write
latency on the disks with separate journaling on SAS 15K disks with a
3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned with
the overall operation apply latency that I have been seeing (1-6ms idle is
normal, but up to 60-170ms for a moderate workload using rbd bench-write)
however I am on a network where I am bound to 1500 mtu and I will get to
test jumbo frames with the next setup in addition to the ssd¹s. I suspect
the overall performance will be good in the new test setup and I am
curious to see what my tests will yield.

Thanks for the response!

Karol



On 2014-03-15, 12:18 PM, "Wido den Hollander"  wrote:

>On 03/15/2014 04:11 PM, Karol Kozubal wrote:
>> Hi Everyone,
>>
>> I am just wondering if any of you are running a ceph cluster with an
>> iSCSI target front end? I know this isn¹t available out of the box,
>> unfortunately in one particular use case we are looking at providing
>> iSCSI access and it's a necessity. I am liking the idea of having rbd
>> devices serving block level storage to the iSCSI Target servers while
>> providing a unified backed for native rbd access by openstack and
>> various application servers. On multiple levels this would reduce the
>> complexity of our SAN environment and move us away from expensive
>> proprietary solutions that don¹t scale out.
>>
>> If any of you have deployed any HA iSCSI Targets backed by rbd I would
>> really appreciate your feedback and any thoughts.
>>
>
>I haven't used it in production, but a couple of things which come to
>mind:
>
>- Use TGT so you can run it all in userspace backed by librbd
>- Do not use writeback caching on the targets
>
>You could use multipathing if you don't use writeback caching. Use
>writeback would also cause data loss/corruption in case of multiple
>targets.
>
>It will probably just work with TGT, but I don't know anything about the
>performance.
>
>> Karol
>>
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>-- 
>Wido den Hollander
>42on B.V.
>
>Phone: +31 (0)20 700 9902
>Skype: contact42on
>___
>ceph-users mailing list
>ceph-users@lists.ceph.com
>http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] RBD as backend for iSCSI SAN Targets

2014-03-15 Thread Wido den Hollander

On 03/15/2014 04:11 PM, Karol Kozubal wrote:

Hi Everyone,

I am just wondering if any of you are running a ceph cluster with an
iSCSI target front end? I know this isn’t available out of the box,
unfortunately in one particular use case we are looking at providing
iSCSI access and it's a necessity. I am liking the idea of having rbd
devices serving block level storage to the iSCSI Target servers while
providing a unified backed for native rbd access by openstack and
various application servers. On multiple levels this would reduce the
complexity of our SAN environment and move us away from expensive
proprietary solutions that don’t scale out.

If any of you have deployed any HA iSCSI Targets backed by rbd I would
really appreciate your feedback and any thoughts.



I haven't used it in production, but a couple of things which come to mind:

- Use TGT so you can run it all in userspace backed by librbd
- Do not use writeback caching on the targets

You could use multipathing if you don't use writeback caching. Use 
writeback would also cause data loss/corruption in case of multiple targets.


It will probably just work with TGT, but I don't know anything about the 
performance.



Karol


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




--
Wido den Hollander
42on B.V.

Phone: +31 (0)20 700 9902
Skype: contact42on
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com