Re: bandwidth with Ceph - v0.59 (Bobtail)

2014-04-25 Thread Xing
running ~10 concurrent workloads. For your case, you were running 256 concurrent read streams for 6/8 disks, I would expect the aggregated disk bandwidth too be lower than 100 MB/s per disk. Any thoughts? Thanks, Xing On Apr 25, 2014, at 5:47 PM, Xing Lin wrote: > Hi Mark, > > T

Re: bandwidth with Ceph - v0.59 (Bobtail)

2014-04-25 Thread Xing Lin
concern is I am not sure how close RADOS bench results are when compared with kernel RBD performance. I would appreciate it if you can do that. Thanks, Xing On 04/25/2014 04:16 PM, Mark Nelson wrote: I don't have any recent results published, but you can see some of the older results from bo

Re: bandwidth with Ceph - v0.59 (Bobtail)

2014-04-25 Thread Xing Lin
Hi Mark, That seems pretty good. What is the block level sequential read bandwidth of your disks? What configuration did you use? What was the replica size, read_ahead for your rbds and what were the number of workloads you used? I used btrfs in my experiments as well. Thanks, Xing On 04

Re: bandwidth with Ceph - v0.59 (Bobtail)

2014-04-25 Thread Xing
. I also measured the maximum bandwidth from each disk we can get in Version 0.79. It does not improve significantly: we can still only get 90~100 MB/s from each disk. Thanks, Xing On Apr 25, 2014, at 2:42 PM, Gregory Farnum wrote: > Bobtail is really too old to draw any meaningful conclusi

bandwidth with Ceph - v0.59 (Bobtail)

2014-04-25 Thread Xing
efficiency is about 60%, what are the reasons that cause this? Could it be because of the locks (futex as I mentioned in my previous email) or anything else? Thanks very much for any feedback. Thanks, Xing -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the

question about futex calls in OSD

2014-04-25 Thread Xing
Hi, This is Xing, a graduate student at the University of Utah. I am playing with Ceph and have a few questions about futex calls in OSD process. I am using Ceph-v0.79. The data node machine has 7 disks, one for OS and the other 6 disks for running 6 OSDs. I set replica size = 1, for all three

Re: unable to compile

2013-11-03 Thread Xing Lin
Thanks, Noah! Xing On 11/3/2013 3:17 PM, Noah Watkins wrote: Thanks for looking at this. Unless there is a good solution I think reverting it is ok as breaking the compile on a few platflorms is not ok. Ill be lookong at this tonight. -- To unsubscribe from this list: send the line

Re: unable to compile

2013-11-03 Thread Xing
Noah. 6efc2b54d5ce85fcb4b66237b051bcbb5072e6a3 Noah, do you have any feedback? Thanks, Xing On Nov 3, 2013, at 1:22 PM, Xing wrote: > Hi all, > > I was able to compile and run vstart.sh yesterday. When I merged the ceph > master branch into my local master branch, I am not abl

unable to compile

2013-11-03 Thread Xing
--enable-plugin --enable-objc-gc --disable-werror --with-arch-32=i686 --with-tune=generic --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 4.6.3 (Ubuntu/Linaro 4.6.3-1ubuntu5) Thanks, Xing -- To unsubscri

Re: [PATCH] test/libcephfs: free cmount after tests finishes

2013-11-02 Thread Xing
Hi Sage, Thanks for applying these two patches. I will try to accumulate more fixes and submit pull requests via github later. Thanks, Xing On Nov 3, 2013, at 12:17 AM, Sage Weil wrote: > Applied this one too! > > BTW, an easier workflow than sending patches to the list is to a

[PATCH] osd/erasurecode: correct one variable name in jerasure_matrix_to_bitmatrix()

2013-11-02 Thread Xing Lin
When bitmatrix is NULL, this function returns NULL. Signed-off-by: Xing Lin --- src/osd/ErasureCodePluginJerasure/jerasure.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/osd/ErasureCodePluginJerasure/jerasure.c b/src/osd/ErasureCodePluginJerasure/jerasure.c index

[PATCH] osd/erasurecode: Fix memory leak in jerasure_matrix_to_bitmatrix()

2013-11-01 Thread Xing Lin
Free allocated memory before return because of NULL input Signed-off-by: Xing Lin --- src/osd/ErasureCodePluginJerasure/jerasure.c | 5 - 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/src/osd/ErasureCodePluginJerasure/jerasure.c b/src/osd/ErasureCodePluginJerasure/jerasure.c

Re: coverity scan - a plea for help!

2013-10-31 Thread Xing Lin
Hi Sage, I would like to help here as well. Thanks, Xing On 10/31/2013 5:30 PM, Sage Weil wrote: Hi everyone, When I send this out several months ago, Danny Al-Gaaf stepped up and submitted an amazing number of patches cleaning up the most concerning issues that Coverity had picked up. His

Re: When ceph synchronizes journal to disk?

2013-03-05 Thread Xing Lin
Thanks very much for all your explanations. I am now much clearer about it. Have a great day! Xing On 03/05/2013 01:12 PM, Greg Farnum wrote: All the data goes to the disk in write-back mode so it isn't safe yet >until the flush is called. That's why it goes into the journal

Re: When ceph synchronizes journal to disk?

2013-03-04 Thread Xing Lin
Maybe it is easier to tell in this way. What we want to see is that the newly written data to stay in the journal disk for as long as possible such that write workloads do not compete for disk headers for read workloads. Any way to achieve that in Ceph? Thanks, Xing On 03/04/2013 09:55 AM

Re: When ceph synchronizes journal to disk?

2013-03-04 Thread Xing Lin
er of enhancements you'll appreciate! Yeah, I would like to use bobtail series. However, I started to make small changes with Argonaut (0.48) and had ported my changes once to 0.48.2 when it was released. I think I am good to continue with it for the moment. I may consider to port my

Re: When ceph synchronizes journal to disk?

2013-03-04 Thread Xing Lin
No, I am using xfs. The same thing happens even if I specified the journal mode explicitly as follows. filestore journal parallel = false filestore journal writeahead = true Xing On 03/04/2013 09:32 AM, Sage Weil wrote: Are you using btrfs? In that case, the journaling is

When ceph synchronizes journal to disk?

2013-03-03 Thread Xing Lin
filestore xattr use omap = true # The maximum interval in seconds for synchronizing the filestore. filestore min sync interval = 60 filestore max sync interval = 300 - Xing -- To unsubscribe from this list: send the line "unsubscribe ceph-devel&

Re: ssh passwords

2013-01-22 Thread Xing Lin
I did not notice that there exists such a parameter. Thanks, Dan! Xing On 01/22/2013 02:11 PM, Dan Mick wrote: The '-a/--allhosts' parameter is to spread the command across the cluster...that is, service ceph -a start will start across the cluster. -- To unsubscribe from this

Re: ssh passwords

2013-01-22 Thread Xing Lin
local osd, mon processes. So, currently, I am using polysh to run the same commands at all hosts (mostly, to restart ceph service before every measurement.). Thanks. Xing On 01/22/2013 12:35 PM, Neil Levine wrote: Out of interest, would people prefer that the Ceph deployment script didn'

Re: ssh passwords

2013-01-22 Thread Xing Lin
osd at other hosts, it will ask us ssh password seven times. So, we'd better set up public key first. :) Xing On 01/22/2013 11:24 AM, Gandalf Corvotempesta wrote: Hi all, i'm trying my very first ceph installation following the 5-minutes quickstart: http://ceph.com/docs/master/sta

Re: Test infrastructure: 2 or more servers?

2013-01-15 Thread Xing Lin
It seems to be: Ceph will shuffle data to rebalance in situations such as when we change the replica num or when some nodes or disks are down. Xing On 01/15/2013 03:26 PM, Gandalf Corvotempesta wrote: So it's absolutely safe to start with just 2 server, make all the necessary tests and

Re: Test infrastructure: 2 or more servers?

2013-01-15 Thread Xing Lin
It seems to be: Ceph will shuffle data to rebalance in situations such as when we change the replica num or when some nodes or disks are down. Xing On 01/15/2013 03:26 PM, Gandalf Corvotempesta wrote: 2013/1/15 Xing Lin : You can change the number replicas in runtime with the following

Re: Test infrastructure: 2 or more servers?

2013-01-15 Thread Xing Lin
You can change the number replicas in runtime with the following command: $ ceph osd pool set {poolname} size {num-replicas} Xing On 01/15/2013 03:00 PM, Gandalf Corvotempesta wrote: Is possible to change the number of replicas in realtime ? -- To unsubscribe from this list: send the line

[debug help]: get dprintk() outputs in src/crush/mapper.c or net/crush/mapper.c

2013-01-13 Thread Xing Lin
crush_do_rule() or my simple bucket_directmap_choose() does not give me any output. Thanks. Xing -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: which Linux kernel version corresponds to 0.48argonaut?

2013-01-05 Thread Xing Lin
It works now. The old version of .so files in /usr/lib are linked, instead of new version of these files which are installed at /usr/local/lib. Thanks, Sage. Xing On 01/05/2013 09:46 PM, Sage Weil wrote: Hi, The rbd binary is dynamically linking to librbd1.so and librados2.so (usually in

Re: which Linux kernel version corresponds to 0.48argonaut?

2013-01-05 Thread Xing Lin
:/users/utos# /usr/bin/rbd -v ceph version 0.48.2argonaut (commit:3e02b2fad88c2a95d9c0c86878f10d1beb780bfe) root@client:/users/utos# /usr/local/bin/rbd -v ceph version 0.48.2argonaut.fast (commit:0) Xing On 01/05/2013 08:00 PM, Mark Kirkwood wrote: I&#

Re: which Linux kernel version corresponds to 0.48argonaut?

2013-01-05 Thread Xing Lin
ad::entry()+0xd) [0x7f1477457c1d] 18: (()+0x7e9a) [0x7f1477152e9a] 19: (clone()+0x6d) [0x7f14759744bd] 2013-01-05 18:29:40.640923 7f14733ca700 -1 *** Caught signal (Aborted) ** in thread 7f14733ca700 Xing On 01/04/2013 05:34 PM, Sage Weil wrote: On Fri, 4 Jan 2013, Gregory Farnum wrote: I th

Re: which Linux kernel version corresponds to 0.48argonaut?

2012-12-20 Thread Xing Lin
implementation. It would be appreciated if someone can confirm it. Thank you! Xing On 12/20/2012 11:54 AM, Xing Lin wrote: Hi, I was trying to add a simple replica placement algorithm in Ceph. This algorithm simply returns r_th item in a bucket for the r_th replica. I have made that change

which Linux kernel version corresponds to 0.48argonaut?

2012-12-20 Thread Xing Lin
ern void crush_destroy_bucket_straw(s extern void crush_destroy_bucket(struct crush_bucket *b); extern void crush_destroy(struct crush_map *map); -static inline int crush_calc_tree_node(int i) -{ -return ((i+1) << 1)-1; -} - #endif Xing -- To unsubscribe from this list: send the line &qu

questions with rbd sequential read throughputs inside kvm/qemu VMs

2012-11-27 Thread Xing Lin
} - Does anyone have some suggestions or hints for me to try? Thank you very much! Xing -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html