Re: [ceph-users] The OSD can be “down” but still “in”.

2019-01-22 Thread M Ranga Swami Reddy
Thanks for reply. If the OSD represents the primary one for a PG, then all IO will be stopped..which may lead to application failure.. On Tue, Jan 22, 2019 at 5:32 PM Matthew Vernon wrote: > > Hi, > > On 22/01/2019 10:02, M Ranga Swami Reddy wrote: > > Hello - If an OSD shown as down and but

Re: [ceph-users] MDS performance issue

2019-01-22 Thread Yan, Zheng
On Wed, Jan 23, 2019 at 10:02 AM Albert Yue wrote: > > But with enough memory on MDS, I can just cache all metadata into memory. > Right now there are around 500GB metadata in the ssd. So this is not enough? > mds needs to tracking lots of extra information for each object. For 500G metadata,

Re: [ceph-users] MDS performance issue

2019-01-22 Thread Albert Yue
But with enough memory on MDS, I can just cache all metadata into memory. Right now there are around 500GB metadata in the ssd. So this is not enough? On Tue, Jan 22, 2019 at 5:48 PM Yan, Zheng wrote: > On Tue, Jan 22, 2019 at 10:49 AM Albert Yue > wrote: > > > > Hi Yan Zheng, > > > > In your

Re: [ceph-users] Broken CephFS stray entries?

2019-01-22 Thread Yan, Zheng
On Tue, Jan 22, 2019 at 10:42 PM Dan van der Ster wrote: > > On Tue, Jan 22, 2019 at 3:33 PM Yan, Zheng wrote: > > > > On Tue, Jan 22, 2019 at 9:08 PM Dan van der Ster > > wrote: > > > > > > Hi Zheng, > > > > > > We also just saw this today and got a bit worried. > > > Should we change to: > >

Re: [ceph-users] cephfs performance degraded very fast

2019-01-22 Thread Yan, Zheng
On Tue, Jan 22, 2019 at 8:24 PM renjianxinlover wrote: > > hi, >at some time, as cache pressure or caps release failure, client apps mount > got stuck. >my use case is in kubernetes cluster and automatic kernel client mount in > nodes. >is anyone faced with same issue or has related

Re: [ceph-users] Process stuck in D+ on cephfs mount

2019-01-22 Thread Yan, Zheng
On Wed, Jan 23, 2019 at 5:50 AM Marc Roos wrote: > > > I got one again > > [] wait_on_page_bit_killable+0x83/0xa0 > [] __lock_page_or_retry+0xb2/0xc0 > [] filemap_fault+0x3b7/0x410 > [] ceph_filemap_fault+0x13c/0x310 [ceph] > [] __do_fault+0x4c/0xc0 > [] do_read_fault.isra.42+0x43/0x130 > []

Re: [ceph-users] Using Ceph central backup storage - Best practice creating pools

2019-01-22 Thread Christian Wuerdig
If you use librados directly it's up to you to ensure you can identify your objects. Generally RADOS stores objects and not files so when you provide your object ids you need to come up with a convention so you can correctly identify them. If you need to provide meta data (i.e. a list of all

Re: [ceph-users] Process stuck in D+ on cephfs mount

2019-01-22 Thread Marc Roos
I got one again [] wait_on_page_bit_killable+0x83/0xa0 [] __lock_page_or_retry+0xb2/0xc0 [] filemap_fault+0x3b7/0x410 [] ceph_filemap_fault+0x13c/0x310 [ceph] [] __do_fault+0x4c/0xc0 [] do_read_fault.isra.42+0x43/0x130 [] handle_mm_fault+0x6b1/0x1040 [] __do_page_fault+0x154/0x450 []

Re: [ceph-users] read-only mounts of RBD images on multiple nodes for parallel reads

2019-01-22 Thread Void Star Nill
Thanks all for the great advices and inputs. Regarding Mykola's suggestion to use Read-Only snapshots, what is the overhead of creating these snapshots? I assume these are copy-on-write snapshots, so there's no extra space consumed except for the metadata? Thanks, Shridhar On Fri, 18 Jan 2019

Re: [ceph-users] Using Ceph central backup storage - Best practice creating pools

2019-01-22 Thread Jack
AFAIK, the only AAA available with librados works on a pool granularity So, if you create a ceph user with access to your pool, he will get access to all the content stored in this pool If you want to use librados for your use case, you will need to implement, on your code, the application logic

Re: [ceph-users] Using Ceph central backup storage - Best practice creating pools

2019-01-22 Thread cmonty14
My backup client is using librados. I understand that defining a pool for the same application is recommended. However this would not answer my other questions: How can I identify a backup created by client A that I want to restore on another client Z? I mean typically client A would write a

[ceph-users] Spec for Ceph Mon+Mgr?

2019-01-22 Thread jesper
Hi. We're currently co-locating our mons with the head node of our Hadoop installation. That may be giving us some problems, we dont know yet, but thus I'm speculation about moving them to dedicated hardware. It is hard to get specifications "small" engough .. the specs for the mon is where we

Re: [ceph-users] Broken CephFS stray entries?

2019-01-22 Thread Dan van der Ster
On Tue, Jan 22, 2019 at 3:33 PM Yan, Zheng wrote: > > On Tue, Jan 22, 2019 at 9:08 PM Dan van der Ster wrote: > > > > Hi Zheng, > > > > We also just saw this today and got a bit worried. > > Should we change to: > > > > What is the error message (on stray dir or other dir)? does the > cluster

Re: [ceph-users] Broken CephFS stray entries?

2019-01-22 Thread Yan, Zheng
On Tue, Jan 22, 2019 at 9:08 PM Dan van der Ster wrote: > > Hi Zheng, > > We also just saw this today and got a bit worried. > Should we change to: > What is the error message (on stray dir or other dir)? does the cluster ever enable multi-acitive mds? > diff --git a/src/mds/CInode.cc

Re: [ceph-users] monitor cephfs mount io's

2019-01-22 Thread Mohamad Gebai
Hi Marc, My point was that there was no way to do that for a kernel mount except from the client that consumes the mounted RBDs. Mohamad On 1/21/19 4:29 AM, Marc Roos wrote: > > Hi Mohamad, How do you do that client side, I am having currently two > kernel mounts? > > > > > > -Original

Re: [ceph-users] Using Ceph central backup storage - Best practice creating pools

2019-01-22 Thread ceph
Hi, Ceph's pool are meant to let you define specific engineering rules and/or application (rbd, cephfs, rgw) They are not designed to be created in a massive fashion (see pgs etc) So, create a pool for each engineering ruleset, and store your data in them For what is left of your project, I

[ceph-users] backfill_toofull while OSDs are not full

2019-01-22 Thread Wido den Hollander
Hi, I've got a couple of PGs which are stuck in backfill_toofull, but none of them are actually full. "up": [ 999, 1900, 145 ], "acting": [ 701, 1146, 1880 ], "backfill_targets": [ "145", "999", "1900" ], "acting_recovery_backfill": [ "145",

Re: [ceph-users] RBD client hangs

2019-01-22 Thread Jason Dillaman
Your "mon" cap should be "profile rbd" instead of "allow r" [1]. [1] http://docs.ceph.com/docs/master/rbd/rados-rbd-cmds/#create-a-block-device-user On Mon, Jan 21, 2019 at 9:05 PM ST Wong (ITSC) wrote: > > Hi, > > > Is this an upgraded or a fresh cluster? > It's a fresh cluster. > > > Does

[ceph-users] Using Ceph central backup storage - Best practice creating pools

2019-01-22 Thread cmonty14
Hi, my use case for Ceph is providing a central backup storage. This means I will backup multiple databases in Ceph storage cluster. This is my question: What is the best practice for creating pools & images? Should I create multiple pools, means one pool per database? Or should I create a

Re: [ceph-users] Broken CephFS stray entries?

2019-01-22 Thread Dan van der Ster
Hi Zheng, We also just saw this today and got a bit worried. Should we change to: diff --git a/src/mds/CInode.cc b/src/mds/CInode.cc index e8c1bc8bc1..e2539390fb 100644 --- a/src/mds/CInode.cc +++ b/src/mds/CInode.cc @@ -2040,7 +2040,7 @@ void CInode::finish_scatter_gather_update(int type)

Re: [ceph-users] migrate ceph-disk to ceph-volume fails with dmcrypt

2019-01-22 Thread Alfredo Deza
On Tue, Jan 22, 2019 at 6:45 AM Manuel Lausch wrote: > > Hi, > > we want upgrade our ceph clusters from jewel to luminous. And also want > to migrate the osds to ceph-volume described in > http://docs.ceph.com/docs/luminous/ceph-volume/simple/scan/#ceph-volume-simple-scan > > The clusters are

[ceph-users] cephfs performance degraded very fast

2019-01-22 Thread renjianxinlover
hi, at some time, as cache pressure or caps release failure, client apps mount got stuck. my use case is in kubernetes cluster and automatic kernel client mount in nodes. is anyone faced with same issue or has related solution? Brs___

Re: [ceph-users] The OSD can be “down” but still “in”.

2019-01-22 Thread Matthew Vernon
Hi, On 22/01/2019 10:02, M Ranga Swami Reddy wrote: > Hello - If an OSD shown as down and but its still "in" state..what > will happen with write/read operations on this down OSD? It depends ;-) In a typical 3-way replicated setup with min_size 2, writes to placement groups on that OSD will

[ceph-users] migrate ceph-disk to ceph-volume fails with dmcrypt

2019-01-22 Thread Manuel Lausch
Hi, we want upgrade our ceph clusters from jewel to luminous. And also want to migrate the osds to ceph-volume described in http://docs.ceph.com/docs/luminous/ceph-volume/simple/scan/#ceph-volume-simple-scan The clusters are running since dumpling and are setup with dmcrypt. Since dumpling there

Re: [ceph-users] Using Ceph central backup storage - Best practice creating pools

2019-01-22 Thread Eugen Block
Hi Thomas, What is the best practice for creating pools & images? Should I create multiple pools, means one pool per database? Or should I create a single pool "backup" and use namespace when writing data in the pool? I don't think one pool per DB is reasonable. If the number of DBs

[ceph-users] The OSD can be “down” but still “in”.

2019-01-22 Thread M Ranga Swami Reddy
Hello - If an OSD shown as down and but its still "in" state..what will happen with write/read operations on this down OSD? Thanks Swami ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] MDS performance issue

2019-01-22 Thread Yan, Zheng
On Tue, Jan 22, 2019 at 10:49 AM Albert Yue wrote: > > Hi Yan Zheng, > > In your opinion, can we resolve this issue by move MDS to a 512GB or 1TB > memory machine? > The problem is from client side, especially clients with large memory. I don't think enlarge mds cache size is good idea. you can

Re: [ceph-users] quick questions about a 5-node homelab setup

2019-01-22 Thread Janne Johansson
Den tis 22 jan. 2019 kl 00:50 skrev Brian Topping : > > I've scrounged up 5 old Atom Supermicro nodes and would like to run them > > 365/7 for limited production as RBD with Bluestore (ideally latest 13.2.4 > > Mimic), triple copy redundancy. Underlying OS is a Debian 9 64 bit, minimal > >

[ceph-users] predict impact of crush tunables change

2019-01-22 Thread Wolfgang Lendl
dear all, i have a luminious cluster with tunables profile "hammer" - now all my hammer clients are gone and i could raise the tunables level to "jewel". is there any good way to predict the data movement caused by such a config change? br wolfgang smime.p7s Description: S/MIME

[ceph-users] Does "mark_unfound_lost delete" only delete missing/unfound objects of a PG

2019-01-22 Thread Mathijs van Veluw
Hello. I have a question about `ceph pg {pg.num} mark_unfound_lost delete`. Will this only delete objects which are unfound, or the whole PG which you put in as an argument? Objects (oid's) which i can see with `ceph pg {pg.num} list_missing`? So in the case bellow, would it remove the object

[ceph-users] krbd reboot hung

2019-01-22 Thread Gao, Wenjun
I’m using krbd to map a rbd device to a VM, it appears when the device is mounted, reboot OS will hung for more than 7min, in baremetal case, it could be more than 15min, even using the latest kernel 5.0.0, the problem still occurs. Here are the console logs with 4.15.18 kernel and mimic rbd

[ceph-users] RadosGW replication and failover issues

2019-01-22 Thread Rom Freiman
Hi, We are running the following radosgw( luminous 12.2.8) replications scenario. 1) We have 2 clusters, each running a radosgw, Cluster1 defined as master, and Cluster2 as slave. 2) We create a number of bucket with objects via master and slave 3) We shutdown the Cluster1 4) We execute failover