Hi Somnath,
Sorry to get your hopes up, no ceph+rocksdb benchmarks (yet!). I was
referring to the benchmarks that the rocksdb developers published here:
https://github.com/facebook/rocksdb/wiki/Performance-Benchmarks
Sounds like we need to start some performance testing on wip-rocksdb
going though! :)
Mark
On 06/05/2014 01:26 PM, Somnath Roy wrote:
Mark,
Could you please share the performance benchmark result with Rocksdbstore +
ceph and leveldbstore+ceph as you mentioned below ?
BTW, have you measured the WA induced by Rocksdbstore and leveldbstore in the
process since that is also a very important factor while backend is flash ?
Thanks & Regards
Somnath
-----Original Message-----
From: [email protected]
[mailto:[email protected]] On Behalf Of Mark Nelson
Sent: Thursday, June 05, 2014 11:18 AM
To: Samuel Just; [email protected]
Subject: Re: Mon backing store
On 06/05/2014 12:42 PM, Samuel Just wrote:
I am starting to wonder whether using leveldb for the mon is actually
introducing an excessive amount unnecessary complexity and
non-determinism. Given that the monitor workload is read mostly,
except for failure conditions when it becomes write latency sensitive,
might we do better with a strict b-tree style backing db such as
berkely db even at the cost of some performance? It seems like
something like that might provide more reliable latency properties.
I'm not against trying it, but I'm not convinced it's the right solution. If
the 99th percentile latency is significantly better, that's obviously a win,
but I think we are indeed going to take a big performance hit overall. I'm
more in favor of trying rocksdb first.
I'm certainly not as well versed in the leveldb interface as you or Joao are,
but it appears much of our code in LevelDBStore would be reusable.
I don't know that rocksdb won't have the same issues that leveldb does, but
the rocksdb developers specifically mention leveldb's bad 99th percentile
latencies as a driver for it's development:
"By contrast, we’ve published the RocksDB benchmark results for server side
workloads on Flash. We also measured the performance of LevelDB on these server-workload
benchmarks and found that RocksDB solidly outperforms LevelDB for these IO bound
workloads. We found that LevelDB’s single-threaded compaction process was insufficient to
drive server workloads. We saw frequent write-stalls with LevelDB that caused
99-percentile latency to be tremendously large. We found that mmap-ing a file into the OS
cache introduced performance bottlenecks for reads. We could not make LevelDB consume all
the IOs offered by the underlying Flash storage."
Compaction performance and high mmap/page fault/kswapd utilization during reads
are two big issues we've hit in leveldb, so I'm inclined to think that rocksdb
is at least worthy of some attention.
Here's the benchmark results on their wiki:
https://github.com/facebook/rocksdb/wiki/Performance-Benchmarks
Thoughts?
-Sam
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel"
in the body of a message to [email protected] More majordomo
info at http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the
body of a message to [email protected] More majordomo info at
http://vger.kernel.org/majordomo-info.html
________________________________
PLEASE NOTE: The information contained in this electronic mail message is
intended only for the use of the designated recipient(s) named above. If the
reader of this message is not the intended recipient, you are hereby notified
that you have received this message in error and that any review,
dissemination, distribution, or copying of this message is strictly prohibited.
If you have received this communication in error, please notify the sender by
telephone or e-mail (as shown above) immediately and destroy any and all copies
of this message in your possession (whether hard copies or electronically
stored copies).
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html