The same..
see:
A: (g=0): rw=read, bs=5M-5M/5M-5M/5M-5M, ioengine=*libaio*, iodepth=1
...
fio-2.2.10
Starting 16 processes

A: (groupid=0, jobs=16): err= 0: pid=27579: Mon Dec 12 20:36:10 2016
  mixed: io=122515MB, bw=6120.3MB/s, iops=1224, runt= 20018msec

I think at the end, the only one way to solve this issue is to write the
image before read test....as suggested....

I have no clue why rbd engine does not work...

On Mon, Dec 12, 2016 at 4:23 PM, Will.Boege <will.bo...@target.com> wrote:

> Try adding --ioengine=libaio
>
>
>
> *From: *V Plus <v.plussh...@gmail.com>
> *Date: *Monday, December 12, 2016 at 2:40 PM
> *To: *"Will.Boege" <will.bo...@target.com>
> *Subject: *Re: [EXTERNAL] [ceph-users] Ceph performance is too good
> (impossible..)...
>
>
>
> Hi Will,
>
> thanks very much..
>
> However, I tried with your suggestions.
>
> Both are *not *working...
>
> 1. with FIO rbd engine:
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> *[RBD_TEST] ioengine=rbd clientname=admin pool=rbd rbdname=fio_test
> invalidate=1    direct=1 group_reporting=1 unified_rw_reporting=1
> time_based=1 rw=read bs=4MB numjobs=16 ramp_time=10 runtime=20*
>
> then I run "sudo fio rbd.job" and  got:
>
>
> RBD_TEST: (g=0): rw=read, bs=4M-4M/4M-4M/4M-4M, ioengine=rbd, iodepth=1
> ...
> fio-2.2.10
> Starting 16 processes
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> rbd engine: RBD version: 0.1.10
> Jobs: 12 (f=7): [R(5),_(1),R(4),_(1),R(2),_(2),R(1)] [100.0% done]
> [11253MB/0KB/0KB /s] [2813/0/0 iops] [eta 00m:00s]
> RBD_TEST: (groupid=0, jobs=16): err= 0: pid=17504: Mon Dec 12 15:32:52 2016
>   mixed: io=212312MB, *bw=10613MB/s*, iops=2653, runt= 20005msec
>
> 2. with blockalign
> [A]
> direct=1
> group_reporting=1
> unified_rw_reporting=1
> size=100%
> time_based=1
> filename=/dev/rbd0
> rw=read
> bs=5MB
> numjobs=16
> ramp_time=5
> runtime=20
> blockalign=512b
>
> [B]
> direct=1
> group_reporting=1
> unified_rw_reporting=1
> size=100%
> time_based=1
> filename=/dev/rbd1
> rw=read
> bs=5MB
> numjobs=16
> ramp_time=5
> runtime=20
> blockalign=512b
>
> sudo fio fioA.job -output a.txt & sudo  fio fioB.job -output b.txt  & wait
>
> Then I got:
> A: (groupid=0, jobs=16): err= 0: pid=19320: Mon Dec 12 15:35:32 2016
>   mixed: io=88590MB, bw=4424.7MB/s, iops=884, runt= 20022msec
>
> B: (groupid=0, jobs=16): err= 0: pid=19324: Mon Dec 12 15:35:32 2016
>   mixed: io=88020MB, *bw=4395.6MB/s*, iops=879, runt= 20025msec
>
> ..............
>
>
>
> On Mon, Dec 12, 2016 at 10:45 AM, Will.Boege <will.bo...@target.com>
> wrote:
>
> My understanding is that when using direct=1 on a raw block device FIO
> (aka-you) will have to handle all the sector alignment or the request will
> get buffered to perform the alignment.
>
>
>
> Try adding the –blockalign=512b option to your jobs, or better yet just
> use the native FIO RBD engine.
>
>
>
> Something like this (untested) -
>
>
>
> [A]
>
> ioengine=rbd
>
> clientname=admin
>
> pool=rbd
>
> rbdname=fio_test
>
> direct=1
>
> group_reporting=1
>
> unified_rw_reporting=1
>
> time_based=1
>
> rw=read
>
> bs=4MB
>
> numjobs=16
>
> ramp_time=10
>
> runtime=20
>
>
>
> *From: *ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of V
> Plus <v.plussh...@gmail.com>
> *Date: *Sunday, December 11, 2016 at 7:44 PM
> *To: *"ceph-users@lists.ceph.com" <ceph-users@lists.ceph.com>
> *Subject: *[EXTERNAL] [ceph-users] Ceph performance is too good
> (impossible..)...
>
>
>
> Hi Guys,
>
> we have a ceph cluster with 6 machines (6 OSD per host).
>
> 1. I created 2 images in Ceph, and map them to another host A (*outside *the
> Ceph cluster). On host A, I got */dev/rbd0* and* /dev/rbd1*.
>
> 2. I start two fio job to perform READ test on rbd0 and rbd1. (fio job
> descriptions can be found below)
>
> *"sudo fio fioA.job -output a.txt & sudo  fio fioB.job -output b.txt  &
> wait"*
>
> 3. After the test, in a.txt, we got *bw=1162.7MB/s*, in b.txt, we get
> *bw=3579.6MB/s*.
>
> The results do NOT make sense because there is only one NIC on host A, and
> its limit is 10 Gbps (1.25GB/s).
>
>
>
> I suspect it is because of the cache setting.
>
> But I am sure that in file */etc/ceph/ceph.conf* on host A,I already
> added:
>
> *[client]*
>
> *rbd cache = false*
>
>
>
> Could anyone give me a hint what is missing? why....
>
> Thank you very much.
>
>
>
> *fioA.job:*
>
> *[A]*
>
> *direct=1*
>
> *group_reporting=1*
>
> *unified_rw_reporting=1*
>
> *size=100%*
>
> *time_based=1*
>
> *filename=/dev/rbd0*
>
> *rw=read*
>
> *bs=4MB*
>
> *numjobs=16*
>
> *ramp_time=10*
>
> *runtime=20*
>
>
>
> *fioB.job:*
>
> *[B]*
>
> *direct=1*
>
> *group_reporting=1*
>
> *unified_rw_reporting=1*
>
> *size=100%*
>
> *time_based=1*
>
> *filename=/dev/rbd1*
>
> *rw=read*
>
> *bs=4MB*
>
> *numjobs=16*
>
> *ramp_time=10*
>
> *runtime=20*
>
>
>
> *Thanks...*
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to