[ceph-users] tgt+librbd error 4

2016-12-16 Thread ZHONG
Hi All, I'm using tgt(1.0.55) + librbd(H 0.94.5) for iSCSI service。Recently encountered problems, TGT in the absence of pressure crush, exception information is as follows:“kernel: tgtd[52067]: segfault at 0 ip 7f424cb0d76a sp 7f4228fe0b90 error 4 in

[ceph-users] list submissions

2018-05-07 Thread ZHONG
___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] ceph-users Digest, Vol 63, Issue 15

2018-04-16 Thread ZHONG
subscribe > 在 2018年4月16日,04:01,ceph-users-requ...@lists.ceph.com 写道: > > Send ceph-users mailing list submissions to > ceph-users@lists.ceph.com > > To subscribe or unsubscribe via the World Wide Web, visit > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > or, via email,

[ceph-users] list submissions

2018-04-16 Thread ZHONG
list submissions ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] machine hangs & soft lockups with 10.2.2 / kernel 4.4.0

2017-01-23 Thread jiajia zhong
try a newer kernel. like 4.8 2017-01-24 0:37 GMT+08:00 Matthew Vernon : > Hi, > > We have a 9-node ceph cluster, running 10.2.2 and kernel 4.4.0 (Ubuntu > Xenial). We're seeing both machines freezing (nothing in logs on the > machine, which is entirely unresponsive to anything

Re: [ceph-users] ceph upgrade from hammer to jewel

2017-02-23 Thread jiajia zhong
are you sure you have ceph-fuse upgraded? #ceph-fuse --version 2017-02-23 16:07 GMT+08:00 gjprabu : > Hi Team, > > We upgraded ceph version from 0.94.9 hammer to 10.2.5 jewel . > Still some clients are showing older version while mounting with debug > mode, is

Re: [ceph-users] Migrate cephfs metadata to SSD in running cluster

2017-02-20 Thread jiajia zhong
his suggested configuration? > > I am kind of a little paranoid to start playing around with crush rules in > the running system. > > Regards, > > Mike > > On 1/5/17 11:40 PM, jiajia zhong wrote: > >> >> >> 2017-01-04 23:52 GMT+08:00 Mike Miller <millerm

Re: [ceph-users] 2 of 3 monitors down and to recover

2017-02-09 Thread jiajia zhong
hi taotao :) you can follow http://docs.ceph.com/docs/master/rados/operations/add-or-rm-mons/ , Remove the non-surviving or problematic monitors. For example ... remember to backup the surviving monitor data before any futher. 2017-02-10 9:52 GMT+08:00 何涛涛(云平台事业部)

[ceph-users] jewel 10.2.5 cephfs fsync write issue

2017-01-18 Thread jiajia zhong
hi guys: I am using kernel client (4.9) to mount cephfs (10.2.5) that just upgraded from Hammer(0.94.9). It became slow while doing fsync write VS same process in Hammer. (Yes, I am sure, fsync is the key.) 1. anyone knows what's goning on ? 2. any way to import that?

[ceph-users] how to get the default CRUSH map that should be generated by ceph itself ?

2016-11-23 Thread JiaJia Zhong
hi, folksIs there any way I could get the original(default) crush map after some manual mofifications? eg: assuming no one had tuned the crush map, the map state named ORIGINAL , which was maintained by ceph itself. if I modified the crush map, map state named COMMIT1,

Re: [ceph-users] cephfs (rbd) read performance low - where is thebottleneck?

2016-11-23 Thread JiaJia Zhong
Mike, if you run mount.ceph with "-v" options, you may get "ceph: Unknown mount option rsize", actually, you could ignore this, the rsize and rasize will both be passed to mount syscall. I belive that you have had the cephfs mounted successfully, run "mount" in terminal to check the actual

[ceph-users] about using SSD in cephfs, attached with some quantified benchmarks

2016-11-25 Thread JiaJia Zhong
confusing questions: (ceph0.94) 1. Is there any way to cache the whole metadata datas into MDS's memory ? (metadata osds dates-async> MDS memory) I dunno if I misunderstand the role of mds :(, so many post threads that advising Using SSD osds for metadata. the metadata stores

Re: [ceph-users] undefined symbol: rados_inconsistent_pg_list

2016-11-28 Thread JiaJia Zhong
hi, 1.try below, remove /root/.python-eggs/rados-0-py2.7-linux-x86_64.egg-tmp/ if you are sure that you want to keep it, back it up. 2. the command you ran, #cp -vf /usr/local/lib/python2.7/site-packages/* /usr/lib64/python2.7/ , generally, this is not recommended. --

Re: [ceph-users] No module named rados

2016-11-28 Thread JiaJia Zhong
hi, since you are playing on centos7, why not following http://docs.ceph.com/docs/master/install/get-packages/ or just downloading the binary packages form https://download.ceph.com/rpm-jewel/ ? :) if you insist to install ceph from ceph-10.2.2.tar.gz, please follow

[ceph-users] can cache-mode be set to readproxy for tier cache with ceph 0.94.9 ?

2016-12-12 Thread JiaJia Zhong
hi cephers: we are using ceph hammer 0.94.9, yes, It's not the latest ( jewel), with some ssd osds for tiering, cache-mode is set to readproxy, everything seems to be as expected, but when reading some small files from cephfs, we got 0 bytes. I did some search and got the

Re: [ceph-users] can cache-mode be set to readproxy for tiercachewith ceph 0.94.9 ?

2016-12-13 Thread JiaJia Zhong
-- Original -- From: "Shinobu Kinjo"<ski...@redhat.com>; Date: Wed, Dec 14, 2016 10:56 AM To: "JiaJia Zhong"<zhongjia...@haomaiyi.com>; Cc: "CEPH list"<ceph-users@lists.ceph.com>; "ukernel"<uker...@

Re: [ceph-users] can cache-mode be set to readproxy for tiercachewith ceph 0.94.9 ?

2016-12-16 Thread JiaJia Zhong
hi skinjo, forgot to ask that if it's necessary to disconnect all the client before doing set-overlay ? we didn't sweep the clients out while setting overlay -- Original -- From: "JiaJia Zhong"<zhongjia...@haomaiyi.com>; Date: Wed, Dec 1

Re: [ceph-users] can cache-mode be set to readproxy for tier cachewith ceph 0.94.9 ?

2016-12-13 Thread JiaJia Zhong
-- Original -- From: "Shinobu Kinjo"<ski...@redhat.com>; Date: Tue, Dec 13, 2016 06:21 PM To: "JiaJia Zhong"<zhongjia...@haomaiyi.com>; Cc: "CEPH list"<ceph-users@lists.ceph.com>; "ukernel"<uker...@

Re: [ceph-users] Ceph performance is too good (impossible..)...

2016-12-11 Thread JiaJia Zhong
>> 3. After the test, in a.txt, we got bw=1162.7MB/s, in b.txt, we get >> bw=3579.6MB/s. mostly, due to your kernel buffer of client host -- Original -- From: "Somnath Roy"; Date: Mon, Dec 12, 2016 09:47 AM To: "V

Re: [ceph-users] 答复: Pipe "deadlock" in Hammer, 0.94.5

2017-01-12 Thread jiajia zhong
f there is pending > msg on the socket by “Pipe::tcp_read_wait”. So, I think, when Pipe::do_recv > is called, it shouldn’t get an EAGAIN, which means it wouldn’t act as > blocked. Is this so? > This really confuses me. > > > 发件人: jiajia zhong [mailto:zhong2p...@gmail.com] > 发送

Re: [ceph-users] Pipe "deadlock" in Hammer, 0.94.5

2017-01-12 Thread jiajia zhong
if errno is EAGAIN for recv, the Pipe:do_recv just acts as blocked. so 2017-01-12 16:34 GMT+08:00 许雪寒 : > Hi, everyone. > > Recently, we did some experiment to test the stability of the ceph > cluster. We used Hammer version which is the mostly used version of online > cluster.

Re: [ceph-users] Analysing ceph performance with SSD journal, 10gbe NIC and 2 replicas -Hammer release

2017-01-05 Thread jiajia zhong
2017-01-06 11:10 GMT+08:00 kevin parrikar : > Hello All, > > I have setup a ceph cluster based on 0.94.6 release in 2 servers each > with 80Gb intel s3510 and 2x3 Tb 7.2 SATA disks,16 CPU,24G RAM > which is connected to a 10G switch with a replica of 2 [ i will add 3

Re: [ceph-users] Migrate cephfs metadata to SSD in running cluster

2017-01-05 Thread jiajia zhong
2017-01-04 23:52 GMT+08:00 Mike Miller : > Wido, all, > > can you point me to the "recent benchmarks" so I can have a look? > How do you define "performance"? I would not expect cephFS throughput to > change, but it is surprising to me that metadata on SSD will have no >

Re: [ceph-users] Ceph - Health and Monitoring

2017-01-04 Thread jiajia zhong
actually, what you need is an ceph-common package (ubuntu) which contains /usr/bin/ceph, You have to be sure the command's going to be executed on which host. make sure the keys and ceph.conf are correctly configured on that host. you could just run the commands to make sure the configure's ok.

Re: [ceph-users] Write back cache removal

2017-01-10 Thread jiajia zhong
It's fixed since v0.94.6, http://ceph.com/releases/v0-94-6-hammer-released/ - fs: CephFS restriction on removing cache tiers is overly strict ( issue#11504 , pr#6402 , John Spray) but you have to make sure

Re: [ceph-users] Mix HDDs and SSDs togheter

2017-03-05 Thread jiajia zhong
we are using mixed too, intel PCIE 400G SSD * 8 for metadata pool and tier caching pool for our cephfs. *plus:* *'osd crush update on start = false*' as Vladimir replied. 2017-03-03 20:33 GMT+08:00 Дробышевский, Владимир : > Hi, Matteo! > > Yes, I'm using mixed cluster in

Re: [ceph-users] Mon stuck in synchronizing after upgrading from Hammer to Jewel

2017-07-04 Thread jiajia zhong
refer to http://docs.ceph.com/docs/master/rados/operations/add-or-rm-mons/ I recalled we encoutered the same issue after upgrading to Jewel :(. 2017-07-05 11:21 GMT+08:00 许雪寒 : > Hi, everyone. > > Recently, we upgraded one of clusters from Hammer to Jewel. However, after >

[ceph-users] is there any way to speed up cache evicting?

2017-06-01 Thread jiajia zhong
hi guys: Our ceph cluster is working with tier cache. I am running "rados -p data_cache cache-try-flush-evict-all" to evict all the objects. But It a bit slow 1. Is there any way to speed up the evicting? 2. Is evicting triggered by itself good enough for cluster ? 3. Does the flushing and

Re: [ceph-users] is there any way to speed up cache evicting?

2017-06-02 Thread jiajia zhong
christian, thanks for your reply. 2017-06-02 11:39 GMT+08:00 Christian Balzer <ch...@gol.com>: > On Fri, 2 Jun 2017 10:30:46 +0800 jiajia zhong wrote: > > > hi guys: > > > > Our ceph cluster is working with tier cache. > If so, then I suppose you read

Re: [ceph-users] is there any way to speed up cache evicting?

2017-06-02 Thread jiajia zhong
thank you for your guide :), It's making sense. 2017-06-02 16:17 GMT+08:00 Christian Balzer <ch...@gol.com>: > > Hello, > > On Fri, 2 Jun 2017 14:30:56 +0800 jiajia zhong wrote: > > > christian, thanks for your reply. > > > > 2017-06-02 11:39 GM

Re: [ceph-users] is there any way to speed up cache evicting?

2017-06-02 Thread jiajia zhong
though cache_target_full_ratio only concerned the clean objects (evict), any misunderstanding ? This fixed our cluster speeds during times when the cache pool was cleaning > up. The problem we hypothesized was that it was the prices of choosing what > to clean vs what to keep was