Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-12 Thread Виталий Филиппов
Hi Felix,

Better use fio.

Like fio -ioengine=rbd -direct=1 -invalidate=1 -name=test -bs=4k -iodepth=128 
-rw=randwrite -pool=rpool_hdd -runtime=60 -rbdname=testimg (for peak parallel 
random iops)

Or the same with -iodepth=1 for the latency test. Here you usually get

Or the same with -ioengine=libaio -filename=testfile -size=10G instead of 
-ioengine=rbd -pool=.. -rbdname=.. to test it from inside a VM.

...or the same with -sync=1 to determine how a DBMS will perform inside a VM...
-- 
With best regards,
  Vitaliy Filippov___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-11 Thread John Petrini
I certainly would, particularly on your SSD's. I'm not familiar with
those Toshibas but disabling disk cache has improved performance on my
clusters and others on this list.

Does the LSI controller you're using provide read/write cache and do
you have it enabled? 72k spinners are likely to see a huge performance
gain from controller cache, especially in regards to latency. Only
enable caching if the controller has a battery and make sure to enable
force write-through in the event that the battery fails. If your
controller doesn't have cache you may want to seriously consider
upgrading to controllers that do otherwise those 72k disks are going
to be a major limiting factor in terms of performance.

Regarding your db partition, the latest advice seems to be that your
db should be 2x the biggest layer (at least 60GB) to avoid spillover
to the OSD during compaction. See:
https://www.mail-archive.com/ceph-users@lists.ceph.com/msg54628.html.
With 72k disks you'll want to avoid small writes hitting them directly
if possible, especially if you have no controller cache.

It would be useful to see iowait on your cluster. iostat -x 2 and let
it run for a few cycles while the cluster is busy. If there's high
iowait on your SSD's disabling disk cache may show an improvement. If
there's high iowait on the HDD's, controller cache and/or increasing
your db size may help.




John Petrini
Platforms Engineer
215.297.4400 x 232
www.coredial.com
751 Arbor Way, Hillcrest I, Suite 150 Blue Bell, PA 19422
The information transmitted is intended only for the person or entity
to which it is addressed and may contain confidential and/or
privileged material. Any review, retransmission, dissemination or
other use of, or taking of any action in reliance upon, this
information by persons or entities other than the intended recipient
is prohibited. If you received this in error, please contact the
sender and delete the material from any computer.


On Tue, Jun 11, 2019 at 3:35 AM Stolte, Felix  wrote:
>
> Hi John,
>
> I have 9 HDDs and 3 SSDs behind a SAS3008 PCI-Express Fusion-MPT SAS-3 from 
> LSI. HDDs are HGST HUH721008AL (8TB, 7200k rpm), SSDs are Toshiba PX05SMB040 
> (400GB). OSDs are bluestore format, 3 HDDs have their wal and db on one SSD 
> (DB Size 50GB, wal 10 GB). I did not change any cache settings.
>
> I disabled cstates which improved performance slightly. Do you suggest to 
> turn off caching on disks?
>
> Regards
> Felix
>
> -
> Forschungszentrum Juelich GmbH
> 52425 Juelich
> Sitz der Gesellschaft: Juelich
> Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> Prof. Dr. Sebastian M. Schmidt
> -
> -
>
>
> Von: John Petrini 
> Datum: Freitag, 7. Juni 2019 um 15:49
> An: "Stolte, Felix" 
> Cc: Sinan Polat , ceph-users 
> Betreff: Re: [ceph-users] Expected IO in luminous Ceph Cluster
>
> How's iowait look on your disks?
>
> How have you configured your disks and what are your cache settings?
>
> Did you disable cstates?
>
> On Friday, June 7, 2019, Stolte, Felix <mailto:f.sto...@fz-juelich.de> wrote:
> > Hi Sinan,
> >
> > thanks for the numbers. I am a little bit surprised that your SSD pool has 
> > nearly the same stats as you SAS pool.
> >
> > Nevertheless I would expect our pools to perform like your SAS pool, at 
> > least regarding to writes since all our write ops should be placed on our 
> > SSDs. But since I only achieve 10% of your numbers I need to figure out my 
> > bottle neck. For now I have no clue. According to our monitoring system 
> > network bandwith, ram or cpu usage is even close to be saturated.
> >
> > Could someone advice me on where to look?
> >
> > Regards Felix
> > -
> > -
> > Forschungszentrum Juelich GmbH
> > 52425 Juelich
> > Sitz der Gesellschaft: Juelich
> > Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> > Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> > Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> > Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> > Prof. Dr. Sebastian M. Schmid

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-11 Thread Stolte, Felix
Hi John,

I have 9 HDDs and 3 SSDs behind a SAS3008 PCI-Express Fusion-MPT SAS-3 from 
LSI. HDDs are HGST HUH721008AL (8TB, 7200k rpm), SSDs are Toshiba PX05SMB040 
(400GB). OSDs are bluestore format, 3 HDDs have their wal and db on one SSD (DB 
Size 50GB, wal 10 GB). I did not change any cache settings. 

I disabled cstates which improved performance slightly. Do you suggest to turn 
off caching on disks?

Regards
Felix

-
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt
-
-


Von: John Petrini 
Datum: Freitag, 7. Juni 2019 um 15:49
An: "Stolte, Felix" 
Cc: Sinan Polat , ceph-users 
Betreff: Re: [ceph-users] Expected IO in luminous Ceph Cluster

How's iowait look on your disks? 

How have you configured your disks and what are your cache settings? 

Did you disable cstates? 

On Friday, June 7, 2019, Stolte, Felix <mailto:f.sto...@fz-juelich.de> wrote:
> Hi Sinan,
>
> thanks for the numbers. I am a little bit surprised that your SSD pool has 
> nearly the same stats as you SAS pool.
>
> Nevertheless I would expect our pools to perform like your SAS pool, at least 
> regarding to writes since all our write ops should be placed on our SSDs. But 
> since I only achieve 10% of your numbers I need to figure out my bottle neck. 
> For now I have no clue. According to our monitoring system network bandwith, 
> ram or cpu usage is even close to be saturated.
>
> Could someone advice me on where to look?
>
> Regards Felix
> -
> -
> Forschungszentrum Juelich GmbH
> 52425 Juelich
> Sitz der Gesellschaft: Juelich
> Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> Prof. Dr. Sebastian M. Schmidt
> -
> -
>
>
> Am 07.06.19, 13:33 schrieb "Sinan Polat" <mailto:si...@turka.nl>:
>
>     Hi Felix,
>
>     I have 2 Pools, a SSD only and a SAS only pool.
>
>     SSD pool is spread over 12 OSD servers.
>     SAS pool is spread over 6 OSD servers.
>
>
>     See results (SSD Only Pool):
>
>     # sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G
>     --file-test-mode=rndrw --file-rw-ratio=2 run
>     sysbench 1.0.17 (using system LuaJIT 2.0.4)
>
>     Running the test with following options:
>     Number of threads: 16
>     Initializing random number generator from current time
>
>
>     Extra file open flags: (none)
>     128 files, 8MiB each
>     1GiB total file size
>     Block size 16KiB
>     Number of IO requests: 0
>     Read/Write ratio for combined random IO test: 2.00
>     Periodic FSYNC enabled, calling fsync() each 1 requests.
>     Calling fsync() at the end of test, Enabled.
>     Using synchronous I/O mode
>     Doing random r/w test
>     Initializing worker threads...
>
>     Threads started!
>
>
>     File operations:
>         reads/s:                      508.38
>         writes/s:                     254.19
>         fsyncs/s:                     32735.14
>
>     Throughput:
>         read, MiB/s:                  7.94
>         written, MiB/s:               3.97
>
>     General statistics:
>         total time:                          10.0103s
>         total number of events:              36
>
>     Latency (ms):
>              min:                                    0.00
>              avg:                                    0.48
>              max:                                   10.18
>              95th percentile:                        2.11
>              sum:                               159830.07
>
>     Threads fairness:
>         events (avg/stddev):           20833.5000/335.70
>         execution time (avg/stddev):   9.9894/0.00
>   

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread John Petrini
:
>  min:0.00
>  avg:0.50
>  max:   20.50
>  95th percentile:2.30
>  sum:   159830.82
>
> Threads fairness:
> events (avg/stddev):   20092.3125/186.66
> execution time (avg/stddev):   9.9894/0.00
> #
>
>
> Kind regards,
> Sinan Polat
>
>
>
> > Op 7 juni 2019 om 12:47 schreef "Stolte, Felix" <
f.sto...@fz-juelich.de>:
> >
> >
> > Hi Sinan,
> >
> > that would be great. The numbers should differ a lot, since you
have an all
> > flash pool, but it would be interesting, what we could expect from
such a
> > configuration.
> >
> > Regards
> > Felix
> >
> >
-
> >
-
> > Forschungszentrum Juelich GmbH
> > 52425 Juelich
> > Sitz der Gesellschaft: Juelich
> > Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> > Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> > Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt
(Vorsitzender),
> > Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> > Prof. Dr. Sebastian M. Schmidt
> >
-
> >
-
> >
> >
> > Am 07.06.19, 12:02 schrieb "Sinan Polat" :
> >
> > Hi Felix,
> >
> > I can run your commands inside an OpenStack VM. Tthe storage
cluster
> > contains of 12 OSD servers, holding each 8x 960GB SSD. Luminous
FileStore.
> > Replicated 3.
> >
> > Would it help you to run your command on my cluster?
> >
> > Sinan
> >
> > > Op 7 jun. 2019 om 08:52 heeft Stolte, Felix <
f.sto...@fz-juelich.de> het
> > > volgende geschreven:
> > >
> > > I have no performance data before we migrated to bluestore.
You should
> > > start a separate topic regarding your question.
> > >
> > > Could anyone with an more or less equally sized cluster post
the output
> > > of a sysbench with the following parameters (either from
inside an
> > > openstack vm or a mounted rbd)?
> > >
> > > sysbench --file-fsync-freq=1 --threads=16 fileio
--file-total-size=1G
> > >--file-test-mode=rndrw --file-rw-ratio=2 prepare
> > >
> > > sysbench --file-fsync-freq=1 --threads=16 fileio
--file-total-size=1G
> > >--file-test-mode=rndrw --file-rw-ratio=2 run
> > >
> > > Thanks in advance.
> > >
> > > Regards
> > > Felix
> > >
> > >
-
> > >
-
> > > Forschungszentrum Juelich GmbH
> > > 52425 Juelich
> > > Sitz der Gesellschaft: Juelich
> > > Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR
B 3498
> > > Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen
Huthmacher
> > > Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt
(Vorsitzender),
> > > Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald
Bolt,
> >     > Prof. Dr. Sebastian M. Schmidt
> > >
-
> > >
-
> > >
> > >
> > > Am 06.06.19, 15:09 schrieb "Marc Roos" <
m.r...@f1-outsourcing.eu>:
> > >
> > >
> > >I am also thinking of moving the wal/db to ssd of the sata
hdd's. Did
> > >
> > >you do tests before and after this change, and know what
the
> > > difference
> > >is 

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Stolte, Felix
rs should differ a lot, since you have an 
all
> flash pool, but it would be interesting, what we could expect from such a
> configuration.
> 
> Regards
> Felix
> 
> 
-
> 
-
> Forschungszentrum Juelich GmbH
> 52425 Juelich
> Sitz der Gesellschaft: Juelich
> Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> Prof. Dr. Sebastian M. Schmidt
> 
-
> 
-
>  
> 
> Am 07.06.19, 12:02 schrieb "Sinan Polat" :
> 
> Hi Felix,
> 
> I can run your commands inside an OpenStack VM. Tthe storage cluster
> contains of 12 OSD servers, holding each 8x 960GB SSD. Luminous FileStore.
> Replicated 3.
> 
> Would it help you to run your command on my cluster?
> 
> Sinan
> 
> > Op 7 jun. 2019 om 08:52 heeft Stolte, Felix 
 het
> > volgende geschreven:
> > 
> > I have no performance data before we migrated to bluestore. You 
should
> > start a separate topic regarding your question.
> > 
> > Could anyone with an more or less equally sized cluster post the 
output
> > of a sysbench with the following parameters (either from inside an
> > openstack vm or a mounted rbd)?
> > 
> > sysbench --file-fsync-freq=1 --threads=16 fileio 
--file-total-size=1G 
> >--file-test-mode=rndrw --file-rw-ratio=2 prepare
> > 
> > sysbench --file-fsync-freq=1 --threads=16 fileio 
--file-total-size=1G 
> >--file-test-mode=rndrw --file-rw-ratio=2 run
> > 
> > Thanks in advance.
> > 
> > Regards
> > Felix
> > 
> > 
-
> > 
-
> > Forschungszentrum Juelich GmbH
> > 52425 Juelich
> > Sitz der Gesellschaft: Juelich
> > Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> > Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> > Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt 
(Vorsitzender),
> > Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> > Prof. Dr. Sebastian M. Schmidt
> > 
-
> > 
-
> > 
> > 
> > Am 06.06.19, 15:09 schrieb "Marc Roos" :
> > 
> > 
> >I am also thinking of moving the wal/db to ssd of the sata 
hdd's. Did
> > 
> >you do tests before and after this change, and know what the
> > difference 
> >is iops? And is the advantage more or less when your sata hdd's 
are 
> >slower? 
> > 
> > 
> >-Original Message-
> >From: Stolte, Felix [mailto:f.sto...@fz-juelich.de] 
> >Sent: donderdag 6 juni 2019 10:47
> >To: ceph-users
> >Subject: [ceph-users] Expected IO in luminous Ceph Cluster
> > 
> >Hello folks,
> > 
> >we are running a ceph cluster on Luminous consisting of 21 OSD 
Nodes 
> >with 9 8TB SATA drives and 3 Intel 3700 SSDs for Bluestore WAL 
and DB
> > 
> >(1:3 Ratio). OSDs have 10Gb for Public and Cluster Network. The
> > cluster 
> >is running stable for over a year. We didn’t had a closer look 
on IO 
> >until one of our customers started to complain about a VM we 
migrated
> > 
> >from VMware with Netapp Storage to our Openstack Cloud with ceph 
> >storage. He sent us a sysbench r

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Sinan Polat
> openstack vm or a mounted rbd)?
> > 
> > sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
> >--file-test-mode=rndrw --file-rw-ratio=2 prepare
> > 
> > sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
> >--file-test-mode=rndrw --file-rw-ratio=2 run
> > 
> > Thanks in advance.
> > 
> > Regards
> > Felix
> > 
> > 
> -
> > 
> -
> > Forschungszentrum Juelich GmbH
> > 52425 Juelich
> > Sitz der Gesellschaft: Juelich
> > Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> > Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> > Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> > Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> > Prof. Dr. Sebastian M. Schmidt
> > 
> -
> > 
> -
> > 
> > 
> > Am 06.06.19, 15:09 schrieb "Marc Roos" :
> > 
> > 
> >I am also thinking of moving the wal/db to ssd of the sata hdd's. Did
> > 
> >you do tests before and after this change, and know what the
> > difference 
> >is iops? And is the advantage more or less when your sata hdd's are 
> >slower? 
> > 
> > 
> >-Original Message-
> >From: Stolte, Felix [mailto:f.sto...@fz-juelich.de] 
> >Sent: donderdag 6 juni 2019 10:47
> >To: ceph-users
> >Subject: [ceph-users] Expected IO in luminous Ceph Cluster
> > 
> >Hello folks,
> > 
> >we are running a ceph cluster on Luminous consisting of 21 OSD Nodes 
> >with 9 8TB SATA drives and 3 Intel 3700 SSDs for Bluestore WAL and DB
> > 
> >(1:3 Ratio). OSDs have 10Gb for Public and Cluster Network. The
> > cluster 
> >is running stable for over a year. We didn’t had a closer look on IO 
> >until one of our customers started to complain about a VM we migrated
> > 
> >from VMware with Netapp Storage to our Openstack Cloud with ceph 
> >storage. He sent us a sysbench report from the machine, which I could
> > 
> >reproduce on other VMs as well as on a mounted RBD on physical
> > hardware:
> > 
> >sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G
> > 
> >--file-test-mode=rndrw --file-rw-ratio=2 run sysbench 1.0.11 (using 
> >system LuaJIT 2.1.0-beta3)
> > 
> >Running the test with following options:
> >Number of threads: 16
> >Initializing random number generator from current time
> > 
> >Extra file open flags: 0
> >128 files, 8MiB each
> >1GiB total file size
> >Block size 16KiB
> >Number of IO requests: 0
> >Read/Write ratio for combined random IO test: 2.00 Periodic FSYNC 
> >enabled, calling fsync() each 1 requests.
> >Calling fsync() at the end of test, Enabled.
> >Using synchronous I/O mode
> >Doing random r/w test
> > 
> >File operations:
> >reads/s:  36.36
> >writes/s: 18.18
> >fsyncs/s: 2318.59
> > 
> >Throughput:
> >read, MiB/s:  0.57
> >written, MiB/s:   0.28
> > 
> >General statistics:
> >total time:  10.0071s
> >total number of events:  23755
> > 
> >Latency (ms):
> > min:  0.01
> > avg:  6.74
> > max:   1112.58
> > 95th percentile: 26.68
> > sum: 160022.67
> > 
> >Threads fairness:
> >events (avg/stddev):   1484.6875/52.59
> >execution time (

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Stolte, Felix
Hi Sinan,

that would be great. The numbers should differ a lot, since you have an all 
flash pool, but it would be interesting, what we could expect from such a 
configuration.

Regards
Felix

-
-
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt
-
-
 

Am 07.06.19, 12:02 schrieb "Sinan Polat" :

Hi Felix,

I can run your commands inside an OpenStack VM. Tthe storage cluster 
contains of 12 OSD servers, holding each 8x 960GB SSD. Luminous FileStore. 
Replicated 3.

Would it help you to run your command on my cluster?

Sinan

> Op 7 jun. 2019 om 08:52 heeft Stolte, Felix  het 
volgende geschreven:
> 
> I have no performance data before we migrated to bluestore. You should 
start a separate topic regarding your question.
> 
> Could anyone with an more or less equally sized cluster post the output 
of a sysbench with the following parameters (either from inside an openstack vm 
or a mounted rbd)?
> 
> sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
>--file-test-mode=rndrw --file-rw-ratio=2 prepare
> 
> sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
>--file-test-mode=rndrw --file-rw-ratio=2 run
> 
> Thanks in advance.
> 
> Regards
> Felix
> 
> 
-
> 
-
> Forschungszentrum Juelich GmbH
> 52425 Juelich
> Sitz der Gesellschaft: Juelich
> Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> Prof. Dr. Sebastian M. Schmidt
> 
-
> 
-
> 
> 
> Am 06.06.19, 15:09 schrieb "Marc Roos" :
> 
> 
>I am also thinking of moving the wal/db to ssd of the sata hdd's. Did 
>you do tests before and after this change, and know what the 
difference 
>is iops? And is the advantage more or less when your sata hdd's are 
>slower? 
> 
> 
>-Original Message-
>    From: Stolte, Felix [mailto:f.sto...@fz-juelich.de] 
>Sent: donderdag 6 juni 2019 10:47
>To: ceph-users
>Subject: [ceph-users] Expected IO in luminous Ceph Cluster
> 
>Hello folks,
> 
>we are running a ceph cluster on Luminous consisting of 21 OSD Nodes 
>with 9 8TB SATA drives and 3 Intel 3700 SSDs for Bluestore WAL and DB 
>(1:3 Ratio). OSDs have 10Gb for Public and Cluster Network. The 
cluster 
>is running stable for over a year. We didn’t had a closer look on IO 
>until one of our customers started to complain about a VM we migrated 
>from VMware with Netapp Storage to our Openstack Cloud with ceph 
>storage. He sent us a sysbench report from the machine, which I could 
>reproduce on other VMs as well as on a mounted RBD on physical 
hardware:
> 
>sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
>--file-test-mode=rndrw --file-rw-ratio=2 run sysbench 1.0.11 (using 
>system LuaJIT 2.1.0-beta3)
> 
>Running the test with following options:
>Number of threads: 16
>Initializing random number generator from current time
> 
>Extra file open flags: 0
>128 files, 8MiB each
>1GiB total file size
>Block size 16KiB
>Number of IO requests: 0
>Read/Write ratio for combined random IO test: 2.00 Periodic FSYNC 
>enabled, calling fsync() each 1 requests.
>Calling fsync() at the end of test, Enabled.
> 

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Sinan Polat
Hi Felix,

I can run your commands inside an OpenStack VM. Tthe storage cluster contains 
of 12 OSD servers, holding each 8x 960GB SSD. Luminous FileStore. Replicated 3.

Would it help you to run your command on my cluster?

Sinan

> Op 7 jun. 2019 om 08:52 heeft Stolte, Felix  het 
> volgende geschreven:
> 
> I have no performance data before we migrated to bluestore. You should start 
> a separate topic regarding your question.
> 
> Could anyone with an more or less equally sized cluster post the output of a 
> sysbench with the following parameters (either from inside an openstack vm or 
> a mounted rbd)?
> 
> sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
>--file-test-mode=rndrw --file-rw-ratio=2 prepare
> 
> sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
>--file-test-mode=rndrw --file-rw-ratio=2 run
> 
> Thanks in advance.
> 
> Regards
> Felix
> 
> -
> -
> Forschungszentrum Juelich GmbH
> 52425 Juelich
> Sitz der Gesellschaft: Juelich
> Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
> Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
> Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
> Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
> Prof. Dr. Sebastian M. Schmidt
> -
> -
> 
> 
> Am 06.06.19, 15:09 schrieb "Marc Roos" :
> 
> 
>I am also thinking of moving the wal/db to ssd of the sata hdd's. Did 
>you do tests before and after this change, and know what the difference 
>is iops? And is the advantage more or less when your sata hdd's are 
>slower? 
> 
> 
>-Original Message-----
>    From: Stolte, Felix [mailto:f.sto...@fz-juelich.de] 
>Sent: donderdag 6 juni 2019 10:47
>To: ceph-users
>Subject: [ceph-users] Expected IO in luminous Ceph Cluster
> 
>Hello folks,
> 
>we are running a ceph cluster on Luminous consisting of 21 OSD Nodes 
>with 9 8TB SATA drives and 3 Intel 3700 SSDs for Bluestore WAL and DB 
>(1:3 Ratio). OSDs have 10Gb for Public and Cluster Network. The cluster 
>is running stable for over a year. We didn’t had a closer look on IO 
>until one of our customers started to complain about a VM we migrated 
>from VMware with Netapp Storage to our Openstack Cloud with ceph 
>storage. He sent us a sysbench report from the machine, which I could 
>reproduce on other VMs as well as on a mounted RBD on physical hardware:
> 
>sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
>--file-test-mode=rndrw --file-rw-ratio=2 run sysbench 1.0.11 (using 
>system LuaJIT 2.1.0-beta3)
> 
>Running the test with following options:
>Number of threads: 16
>Initializing random number generator from current time
> 
>Extra file open flags: 0
>128 files, 8MiB each
>1GiB total file size
>Block size 16KiB
>Number of IO requests: 0
>Read/Write ratio for combined random IO test: 2.00 Periodic FSYNC 
>enabled, calling fsync() each 1 requests.
>Calling fsync() at the end of test, Enabled.
>Using synchronous I/O mode
>Doing random r/w test
> 
>File operations:
>reads/s:  36.36
>writes/s: 18.18
>fsyncs/s: 2318.59
> 
>Throughput:
>read, MiB/s:  0.57
>written, MiB/s:   0.28
> 
>General statistics:
>total time:  10.0071s
>total number of events:  23755
> 
>Latency (ms):
> min:  0.01
> avg:  6.74
> max:   1112.58
> 95th percentile: 26.68
> sum: 160022.67
> 
>Threads fairness:
>events (avg/stddev):   1484.6875/52.59
>execution time (avg/stddev):   10.0014/0.00
> 
>Are these numbers reasonable for a cluster of our size?
> 
>Best regards
>Felix
>IT-Services
>Telefon 02461 61-9243
>E-Mail: f.sto...@fz-juelich.de
>
>-
>--

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-07 Thread Stolte, Felix
I have no performance data before we migrated to bluestore. You should start a 
separate topic regarding your question.

Could anyone with an more or less equally sized cluster post the output of a 
sysbench with the following parameters (either from inside an openstack vm or a 
mounted rbd)?

sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
--file-test-mode=rndrw --file-rw-ratio=2 prepare

sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
--file-test-mode=rndrw --file-rw-ratio=2 run

Thanks in advance.

Regards
Felix

-
-
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt
-
-
 

Am 06.06.19, 15:09 schrieb "Marc Roos" :


I am also thinking of moving the wal/db to ssd of the sata hdd's. Did 
you do tests before and after this change, and know what the difference 
is iops? And is the advantage more or less when your sata hdd's are 
slower? 


-Original Message-
From: Stolte, Felix [mailto:f.sto...@fz-juelich.de] 
Sent: donderdag 6 juni 2019 10:47
To: ceph-users
Subject: [ceph-users] Expected IO in luminous Ceph Cluster

Hello folks,

we are running a ceph cluster on Luminous consisting of 21 OSD Nodes 
with 9 8TB SATA drives and 3 Intel 3700 SSDs for Bluestore WAL and DB 
(1:3 Ratio). OSDs have 10Gb for Public and Cluster Network. The cluster 
is running stable for over a year. We didn’t had a closer look on IO 
until one of our customers started to complain about a VM we migrated 
from VMware with Netapp Storage to our Openstack Cloud with ceph 
storage. He sent us a sysbench report from the machine, which I could 
reproduce on other VMs as well as on a mounted RBD on physical hardware:

sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
--file-test-mode=rndrw --file-rw-ratio=2 run sysbench 1.0.11 (using 
system LuaJIT 2.1.0-beta3)

Running the test with following options:
Number of threads: 16
Initializing random number generator from current time

Extra file open flags: 0
128 files, 8MiB each
1GiB total file size
Block size 16KiB
Number of IO requests: 0
Read/Write ratio for combined random IO test: 2.00 Periodic FSYNC 
enabled, calling fsync() each 1 requests.
Calling fsync() at the end of test, Enabled.
Using synchronous I/O mode
Doing random r/w test

File operations:
reads/s:  36.36
writes/s: 18.18
fsyncs/s: 2318.59

Throughput:
read, MiB/s:  0.57
written, MiB/s:   0.28

General statistics:
total time:  10.0071s
total number of events:  23755

Latency (ms):
 min:  0.01
 avg:  6.74
 max:   1112.58
 95th percentile: 26.68
 sum: 160022.67

Threads fairness:
events (avg/stddev):   1484.6875/52.59
execution time (avg/stddev):   10.0014/0.00

Are these numbers reasonable for a cluster of our size?

Best regards
Felix
IT-Services
Telefon 02461 61-9243
E-Mail: f.sto...@fz-juelich.de

-

-
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498 
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender), 
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt, Prof. 
Dr. Sebastian M. Schmidt

-

-
 

___
ceph-use

Re: [ceph-users] Expected IO in luminous Ceph Cluster

2019-06-06 Thread Marc Roos

I am also thinking of moving the wal/db to ssd of the sata hdd's. Did 
you do tests before and after this change, and know what the difference 
is iops? And is the advantage more or less when your sata hdd's are 
slower? 


-Original Message-
From: Stolte, Felix [mailto:f.sto...@fz-juelich.de] 
Sent: donderdag 6 juni 2019 10:47
To: ceph-users
Subject: [ceph-users] Expected IO in luminous Ceph Cluster

Hello folks,

we are running a ceph cluster on Luminous consisting of 21 OSD Nodes 
with 9 8TB SATA drives and 3 Intel 3700 SSDs for Bluestore WAL and DB 
(1:3 Ratio). OSDs have 10Gb for Public and Cluster Network. The cluster 
is running stable for over a year. We didn’t had a closer look on IO 
until one of our customers started to complain about a VM we migrated 
from VMware with Netapp Storage to our Openstack Cloud with ceph 
storage. He sent us a sysbench report from the machine, which I could 
reproduce on other VMs as well as on a mounted RBD on physical hardware:

sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
--file-test-mode=rndrw --file-rw-ratio=2 run sysbench 1.0.11 (using 
system LuaJIT 2.1.0-beta3)

Running the test with following options:
Number of threads: 16
Initializing random number generator from current time

Extra file open flags: 0
128 files, 8MiB each
1GiB total file size
Block size 16KiB
Number of IO requests: 0
Read/Write ratio for combined random IO test: 2.00 Periodic FSYNC 
enabled, calling fsync() each 1 requests.
Calling fsync() at the end of test, Enabled.
Using synchronous I/O mode
Doing random r/w test

File operations:
reads/s:  36.36
writes/s: 18.18
fsyncs/s: 2318.59

Throughput:
read, MiB/s:  0.57
written, MiB/s:   0.28

General statistics:
total time:  10.0071s
total number of events:  23755

Latency (ms):
 min:  0.01
 avg:  6.74
 max:   1112.58
 95th percentile: 26.68
 sum: 160022.67

Threads fairness:
events (avg/stddev):   1484.6875/52.59
execution time (avg/stddev):   10.0014/0.00

Are these numbers reasonable for a cluster of our size?

Best regards
Felix
IT-Services
Telefon 02461 61-9243
E-Mail: f.sto...@fz-juelich.de

-

-
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498 
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender), 
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt, Prof. 
Dr. Sebastian M. Schmidt

-

-
 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Expected IO in luminous Ceph Cluster

2019-06-06 Thread Stolte, Felix
Hello folks,

we are running a ceph cluster on Luminous consisting of 21 OSD Nodes with 9 8TB 
SATA drives and 3 Intel 3700 SSDs for Bluestore WAL and DB (1:3 Ratio). OSDs 
have 10Gb for Public and Cluster Network. The cluster is running stable for 
over a year. We didn’t had a closer look on IO until one of our customers 
started to complain about a VM we migrated from VMware with Netapp Storage to 
our Openstack Cloud with ceph storage. He sent us a sysbench report from the 
machine, which I could reproduce on other VMs as well as on a mounted RBD on 
physical hardware:

sysbench --file-fsync-freq=1 --threads=16 fileio --file-total-size=1G 
--file-test-mode=rndrw --file-rw-ratio=2 run
sysbench 1.0.11 (using system LuaJIT 2.1.0-beta3)

Running the test with following options:
Number of threads: 16
Initializing random number generator from current time

Extra file open flags: 0
128 files, 8MiB each
1GiB total file size
Block size 16KiB
Number of IO requests: 0
Read/Write ratio for combined random IO test: 2.00
Periodic FSYNC enabled, calling fsync() each 1 requests.
Calling fsync() at the end of test, Enabled.
Using synchronous I/O mode
Doing random r/w test

File operations:
reads/s:  36.36
writes/s: 18.18
fsyncs/s: 2318.59

Throughput:
read, MiB/s:  0.57
written, MiB/s:   0.28

General statistics:
total time:  10.0071s
total number of events:  23755

Latency (ms):
 min:  0.01
 avg:  6.74
 max:   1112.58
 95th percentile: 26.68
 sum: 160022.67

Threads fairness:
events (avg/stddev):   1484.6875/52.59
execution time (avg/stddev):   10.0014/0.00

Are these numbers reasonable for a cluster of our size?

Best regards
Felix
IT-Services
Telefon 02461 61-9243
E-Mail: f.sto...@fz-juelich.de
-
-
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt
-
-
 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com