Re: [ceph-users] cephfs kernel client instability

2019-01-25 Thread Ilya Dryomov
On Fri, Jan 25, 2019 at 9:40 AM Martin Palma wrote: > > > Do you see them repeating every 30 seconds? > > yes: > > Jan 25 09:34:37 sdccgw01 kernel: [6306813.737615] libceph: mon4 > 10.8.55.203:6789 session lost, hunting for new mon > Jan 25 09:34:37 sdccgw01 kernel: [6306813.737620] libceph: mon3

[ceph-users] Docubetter: New Schedule

2019-01-25 Thread Neha Ojha
Hi All, Starting February, Docubetter meetings will be held twice a month, on the second and fourth Wednesday of each month. We will alternate meeting times to ensure that all time zones have the opportunity to participate. Second Wednesday: 12:30 ET (starting February 13) Fourth Wednesday:

Re: [ceph-users] bluestore block.db

2019-01-25 Thread Igor Fedotov
Hi Frank, you might want to use ceph-kvstore-tool, e.g. ceph-kvstore-tool bluestore-kv dump Thanks, Igor On 1/25/2019 10:49 PM, F Ritchie wrote: Hi all, Is there a way to dump the contents of block.db to a text file? I am not trying to fix a problem just curious and want to poke around.

Re: [ceph-users] Creating bootstrap keys

2019-01-25 Thread Randall Smith
I figured it out. I had to dig through the ceph-deploy code to get it but here it is. Apparently, one of the upgrades create the keys in ceph but the bootstrap keyrings were never created. ceph auth get client.bootstrap-osd > /var/lib/ceph/bootstrap-osd/ceph.keyring On Thu, Jan 24, 2019 at 4:56

[ceph-users] bluestore block.db

2019-01-25 Thread F Ritchie
Hi all, Is there a way to dump the contents of block.db to a text file? I am not trying to fix a problem just curious and want to poke around. thx Frank ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Does "mark_unfound_lost delete" only delete missing/unfound objects of a PG

2019-01-25 Thread Mathijs van Veluw
It has been resolved. It seems that it in fact only removes the objects which you list with the list_missing option. On Tue, Jan 22, 2019 at 9:42 AM Mathijs van Veluw < mathijs.van.ve...@gmail.com> wrote: > Hello. > I have a question about `ceph pg {pg.num} mark_unfound_lost delete`. > Will

Re: [ceph-users] ceph osd commit latency increase over time, until restart

2019-01-25 Thread Alexandre DERUMIER
also, here the result of "perf diff 1mslatency.perfdata 3mslatency.perfdata" http://odisoweb1.odiso.net/perf_diff_ok_vs_bad.txt - Mail original - De: "aderumier" À: "Sage Weil" Cc: "ceph-users" , "ceph-devel" Envoyé: Vendredi 25 Janvier 2019 17:32:02 Objet: Re: [ceph-users] ceph

Re: [ceph-users] ceph osd commit latency increase over time, until restart

2019-01-25 Thread Alexandre DERUMIER
Hi again, I was able to perf it today, before restart, commit latency was between 3-5ms after restart at 17:11, latency is around 1ms http://odisoweb1.odiso.net/osd3_latency_3ms_vs_1ms.png here some perf reports: with 3ms latency: - perf report by caller:

Re: [ceph-users] backfill_toofull while OSDs are not full

2019-01-25 Thread Paul Emmerich
I've also seen this effect a few times since Mimic (never happened in Luminous). It always resolved itself but the HEALTH_ERROR can be confusing to users. Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München

Re: [ceph-users] Radosgw s3 subuser permissions

2019-01-25 Thread Adam C. Emerson
On 24/01/2019, Marc Roos wrote: > > > This should do it sort of. > > { > "Id": "Policy1548367105316", > "Version": "2012-10-17", > "Statement": [ > { > "Sid": "Stmt1548367099807", > "Effect": "Allow", > "Action": "s3:ListBucket", > "Principal": { "AWS":

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Jason Dillaman
Your caps are invalid -- you are missing the access-spec on the first OSD cap clause. See [1] for more information for how to properly format caps. What version of Ceph are you using? If I remember correctly, since Luminous it will at least throw an error when you give it invalid caps.

Re: [ceph-users] Modify ceph.mon network required

2019-01-25 Thread Thomas
Thanks. This procedure works very well. Am 25.01.2019 um 14:24 schrieb Janne Johansson: > Den fre 25 jan. 2019 kl 09:52 skrev cmonty14 <74cmo...@gmail.com>: >> Hi, >> I have identified a major issue with my cluster setup consisting of 3 nodes: >> all monitors are connected to cluster network.

Re: [ceph-users] RBD client hangs

2019-01-25 Thread Jason Dillaman
That doesn't appear to be an error -- that's just stating that it found a dead client that was holding the exclusice-lock, so it broke the dead client's lock on the image (by blacklisting the client). On Fri, Jan 25, 2019 at 5:09 AM ST Wong (ITSC) wrote: > > Oops, while I can map and mount the

Re: [ceph-users] krbd reboot hung

2019-01-25 Thread Jason Dillaman
The "rbdmap" systemd unit file should take care of it [1]. [1] https://github.com/ceph/ceph/blob/master/systemd/rbdmap.service.in#L4 On Fri, Jan 25, 2019 at 3:00 AM Gao, Wenjun wrote: > > Thanks, what’s the configuration you mentioned? > > > > -- > > Thanks, > > Wenjun > > > > From: Gregory

Re: [ceph-users] Modify ceph.mon network required

2019-01-25 Thread Janne Johansson
Den fre 25 jan. 2019 kl 09:52 skrev cmonty14 <74cmo...@gmail.com>: > > Hi, > I have identified a major issue with my cluster setup consisting of 3 nodes: > all monitors are connected to cluster network. > > Question: > How can I modify the network configuration of mon? > > It's not working to

Re: [ceph-users] [Solved]reating a block device user with restricted access to image

2019-01-25 Thread Thomas
Update: I have identified the root cause: user caps are not correct. Errornous caps: root@ld4257:/etc/ceph# ceph auth get client.gbsadm exported keyring for client.gbsadm [client.gbsadm]     key = AQBd0klcFknvMRAAwuu30bNG7L7PHk5d8cSVvg==     caps mon = "allow r"     caps osd = "allow

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Thomas
Hi, unfortunately it's not working, yet. I have modified user gbsadm: root@ld4257:/etc/ceph# ceph auth get client.gbsadm exported keyring for client.gbsadm [client.gbsadm]     key = AQBd0klcFknvMRAAwuu30bNG7L7PHk5d8cSVvg==     caps mon = "allow r"     caps osd = "allow pool backup

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Burkhard Linke
Hi, just a comment: the RBD pool also contain management objects, e.g. the rbd_directory and rbd_info objects. And AFAIK these objects storage the name->id mapping for images. This means in your case, looking up the name backup/gbs requires read access to these objects in the backup

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Eugen Block
You can check all objects of that pool to see if your caps match: rados -p backup ls | grep rbd_id Zitat von Eugen Block : caps osd = "allow pool backup object_prefix rbd_data.18102d6b8b4567; allow rwx pool backup object_prefix rbd_header.18102d6b8b4567; allow rx pool backup object_prefix

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Eugen Block
caps osd = "allow pool backup object_prefix rbd_data.18102d6b8b4567; allow rwx pool backup object_prefix rbd_header.18102d6b8b4567; allow rx pool backup object_prefix rbd_id.rbd-image" I think your caps are not entirely correct, the part "[...] object_prefix rbd_id.rbd-image" should contain

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Thomas
Thanks. Unfortunately this is still not working. Here's the info of my image: root@ld4257:/etc/ceph# rbd info backup/gbs rbd image 'gbs':     size 500GiB in 128000 objects     order 22 (4MiB objects)     block_name_prefix: rbd_data.18102d6b8b4567     format: 2     features:

Re: [ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread Eugen Block
Hi, I replied to your thread a couple of days ago, maybe you didn't notice: Restricting user access is possible on rbd image level. You can grant read/write access for one client and only read access for other clients, you have to create different clients for that, see [1] for more

[ceph-users] Creating a block device user with restricted access to image

2019-01-25 Thread cmonty14
Hi, I can create a block device user with this command: ceph auth get-or-create client.{ID} mon 'profile rbd' osd 'profile {profile name} [pool={pool-name}][, profile ...]' Question: How can I create a user that has access only to a specific image created in pool ? If this is not possible this

Re: [ceph-users] RBD client hangs

2019-01-25 Thread ST Wong (ITSC)
Oops, while I can map and mount the filesystem, still found error as below, while rebooting the client machine freezes and have to power reset her. Jan 25 17:57:30 acapp1 kernel: XFS (rbd0): Mounting V5 Filesystem Jan 25 17:57:30 acapp1 kernel: rbd: rbd0: client74700 seems dead, breaking lock

Re: [ceph-users] ceph osd commit latency increase over time, until restart

2019-01-25 Thread Alexandre DERUMIER
>>Can you capture a perf top or perf record to see where teh CPU time is >>going on one of the OSDs wth a high latency? Yes, sure. I'll do it next week and send result to the mailing list. Thanks Sage ! - Mail original - De: "Sage Weil" À: "aderumier" Cc: "ceph-users" , "ceph-devel"

Re: [ceph-users] RBD client hangs

2019-01-25 Thread ST Wong (ITSC)
Hi, It works. Thanks a lot. /st -Original Message- From: Jason Dillaman Sent: Tuesday, January 22, 2019 9:29 PM To: ST Wong (ITSC) Cc: Ilya Dryomov ; ceph-users@lists.ceph.com Subject: Re: [ceph-users] RBD client hangs Your "mon" cap should be "profile rbd" instead of "allow r"

Re: [ceph-users] ceph osd commit latency increase over time, until restart

2019-01-25 Thread Sage Weil
Can you capture a perf top or perf record to see where teh CPU time is going on one of the OSDs wth a high latency? Thanks! sage On Fri, 25 Jan 2019, Alexandre DERUMIER wrote: > > Hi, > > I have a strange behaviour of my osd, on multiple clusters, > > All cluster are running mimic

[ceph-users] ceph osd commit latency increase over time, until restart

2019-01-25 Thread Alexandre DERUMIER
Hi, I have a strange behaviour of my osd, on multiple clusters, All cluster are running mimic 13.2.1,bluestore, with ssd or nvme drivers, workload is rbd only, with qemu-kvm vms running with librbd + snapshot/rbd export-diff/snapshotdelete each day for backup When the osd are refreshly

[ceph-users] Modify ceph.mon network required

2019-01-25 Thread cmonty14
Hi, I have identified a major issue with my cluster setup consisting of 3 nodes: all monitors are connected to cluster network. Here's my /etc/ceph/ceph.conf: [global] auth client required = cephx auth cluster required = cephx auth service required = cephx

Re: [ceph-users] cephfs kernel client instability

2019-01-25 Thread Martin Palma
> Do you see them repeating every 30 seconds? yes: Jan 25 09:34:37 sdccgw01 kernel: [6306813.737615] libceph: mon4 10.8.55.203:6789 session lost, hunting for new mon Jan 25 09:34:37 sdccgw01 kernel: [6306813.737620] libceph: mon3 10.8.55.202:6789 session lost, hunting for new mon Jan 25 09:34:37

Re: [ceph-users] cephfs kernel client instability

2019-01-25 Thread Ilya Dryomov
On Fri, Jan 25, 2019 at 8:37 AM Martin Palma wrote: > > Hi Ilya, > > thank you for the clarification. After setting the > "osd_map_messages_max" to 10 the io errors and the MDS error > "MDS_CLIENT_LATE_RELEASE" are gone. > > The messages of "mon session lost, hunting for new new mon" didn't go >

Re: [ceph-users] krbd reboot hung

2019-01-25 Thread Gao, Wenjun
Thanks, what’s the configuration you mentioned? -- Thanks, Wenjun From: Gregory Farnum Date: Friday, January 25, 2019 at 3:35 PM To: "Gao, Wenjun" Cc: "ceph-users@lists.ceph.com" Subject: Re: [ceph-users] krbd reboot hung Looks like your network deactivated before the rbd volume was