On 22/05/2015 16:25, Barclay Jameson wrote:
The Bonnie++ job _FINALLY_ finished. If I am reading this correctly it
took days to create, stat, and delete 16 files??
[root@blarg cephfs]# ~/bonnie++-1.03e/bonnie++ -u root:root -s 256g -r
131072 -d /cephfs/ -m CephBench -f -b
Using uid:0, gid:0.
Writing intelligently...done
Rewriting...done
Reading intelligently...done
start 'em...done...done...done...
Create files in sequential order...done.
Stat files in sequential order...done.
Delete files in sequential order...done.
Create files in random order...done.
Stat files in random order...done.
Delete files in random order...done.
Version 1.03e       ------Sequential Output------ --Sequential Input- --Random-
                     -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks--
Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec %CP
CephBench      256G           1006417  76 90114  13           137110
8 329.8   7
                     ------Sequential Create------ --------Random Create--------
                     -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete--
               files  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP
                  16     0   0 +++++ +++     0   0     0   0  5267  19     0   0
CephBench,256G,,,1006417,76,90114,13,,,137110,8,329.8,7,16,0,0,+++++,+++,0,0,0,0,5267,19,0,0

Any thoughts?

It's 16000 files by default (not 16), but this usually takes only a few minutes.

FWIW I tried running a quick bonnie++ (with -s 0 to skip the IO phase) on a development (vstart.sh) cluster with a fuse client, and it readily handles several hundred client requests per second (checked with "ceph daemonperf mds.<id>")

Nothing immediately leapt out at me from a quick look at the log you posted, but with issues like these it is always worth trying to narrow it down by trying the fuse client instead of the kernel client, and/or different kernel versions.

You may also want to check that your underlying RADOS cluster is performing reasonably by doing a rados bench too.

Cheers,
John
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to