Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock

2018-03-12 Thread David Disseldorp
On Fri, 09 Mar 2018 11:23:02 +0200, Maged Mokhtar wrote: > 2)I undertand that before switching the path, the initiator will send a > TMF ABORT can we pass this to down to the same abort_request() function > in osd_client that is used for osd_request_timeout expiry ? IIUC, the existing

Re: [ceph-users] XFS Metadata corruption while activating OSD

2018-03-12 Thread Peter Woodman
from what i've heard, xfs has problems on arm. use btrfs, or (i believe?) ext4+bluestore will work. On Sun, Mar 11, 2018 at 9:49 PM, Christian Wuerdig wrote: > Hm, so you're running OSD nodes with 2GB of RAM and 2x10TB = 20TB of > storage? Literally everything posted

Re: [ceph-users] ceph-mds suicide on upgrade

2018-03-12 Thread Dietmar Rieder
Hi, See: http://lists.ceph.com/pipermail/ceph-users-ceph.com/2018-February/025092.html Might be of interest. Dietmar Am 12. März 2018 18:19:51 MEZ schrieb Reed Dier : >Figured I would see if anyone has seen this or can see something I am >doing wrong. > >Upgrading all

Re: [ceph-users] ceph-mds suicide on upgrade

2018-03-12 Thread Reed Dier
Good eye, Thanks Dietmar, Glad to know this isn’t a standard issue, hopefully anything in the future will get caught and/or make it into release notes. Thanks, Reed > On Mar 12, 2018, at 12:55 PM, Dietmar Rieder > wrote: > > Hi, > > See: >

[ceph-users] ceph mount nofail option

2018-03-12 Thread Kenneth Waegeman
Hi all, Is there a way to mount ceph kernel client with the nofail option ? I get an invalid argument when trying to mount ceph with nofail option, in fstab / mount mon01,mon02,mon03:/ /mnt/ceph ceph name=cephfs,secretfile=/etc/ceph/secret,noatime,nofail 0 0 or [root@osd003 ~]# mount -t ceph

[ceph-users] Issue with fstrim and Nova hw_disk_discard=unmap

2018-03-12 Thread Fulvio Galeazzi
Hallo all, I am not sure RBD discard is working in my setup, and I am asking for your help. (I searched this mailing list for related messages and found one by Nathan Harper last 29th Jan 2018 "Debugging fstrim issues" which however mentions trimming was masked by logging... so

Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock

2018-03-12 Thread Maged Mokhtar
On 2018-03-12 14:23, David Disseldorp wrote: > On Fri, 09 Mar 2018 11:23:02 +0200, Maged Mokhtar wrote: > >> 2)I undertand that before switching the path, the initiator will send a >> TMF ABORT can we pass this to down to the same abort_request() function >> in osd_client that is used for

Re: [ceph-users] Fwd: [ceph bad performance], can't find a bottleneck

2018-03-12 Thread Maged Mokhtar
Hi, Try increasing the queue depth from default 128 to 1024: rbd map image-XX -o queue_depth=1024 Also if you run multiple rbd images/fio tests, do you get higher combined performance ? Maged On 2018-03-12 17:16, Sergey Kotov wrote: > Dear moderator, i subscribed to ceph list today,

Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock

2018-03-12 Thread David Disseldorp
Hi Maged, On Mon, 12 Mar 2018 20:41:22 +0200, Maged Mokhtar wrote: > I was thinking we would get the block request then loop down to all its > osd requests and cancel those using the same osd request cancel > function. Until we can be certain of termination, I don't think it makes sense to

Re: [ceph-users] Fwd: Slow requests troubleshooting in Luminous - details missing

2018-03-12 Thread Alex Gorbachev
On Mon, Mar 12, 2018 at 7:53 AM, Дробышевский, Владимир wrote: > > I was following this conversation on tracker and got the same question. I've > got a situation with slow requests and had no any idea on how to find the > reason. Finally I found it but only because I knew I've

[ceph-users] Fwd: [ceph bad performance], can't find a bottleneck

2018-03-12 Thread Sergey Kotov
Dear moderator, i subscribed to ceph list today, could you please post my message? -- Forwarded message -- From: Sergey Kotov Date: 2018-03-06 10:52 GMT+03:00 Subject: [ceph bad performance], can't find a bottleneck To: ceph-users@lists.ceph.com Cc: Житенев

[ceph-users] RGW bucket notifications

2018-03-12 Thread Alex Sainer
Hi All, how do you handle bucket-notifications? Is there any known piece of software we can put in front of ceph-rgw? or anything else? best regards, Alex ___ ceph-users mailing list ceph-users@lists.ceph.com

[ceph-users] ceph-mds suicide on upgrade

2018-03-12 Thread Reed Dier
Figured I would see if anyone has seen this or can see something I am doing wrong. Upgrading all of my daemons from 12.2.2. to 12.2.4. Followed the documentation, upgraded mons, mgrs, osds, then mds’s in that order. All was fine, until the MDSs. I have two MDS’s in Active:Standby config. I

[ceph-users] Fwd: Slow requests troubleshooting in Luminous - details missing

2018-03-12 Thread Дробышевский , Владимир
I was following this conversation on tracker and got the same question. I've got a situation with slow requests and had no any idea on how to find the reason. Finally I found it but only because I knew I've upgraded Mellanox drivers on one host, and just decided to check IB config (and the root

Re: [ceph-users] Issue with fstrim and Nova hw_disk_discard=unmap

2018-03-12 Thread Jason Dillaman
On Mon, Mar 12, 2018 at 9:54 AM, Fulvio Galeazzi wrote: > Hallo all, > I am not sure RBD discard is working in my setup, and I am asking for > your help. > (I searched this mailing list for related messages and found one by > Nathan Harper last 29th Jan 2018

Re: [ceph-users] Civetweb log format

2018-03-12 Thread Aaron Bassett
Quick update: adding the following to your config: rgw log http headers = "http_authorization" rgw ops log socket path = /tmp/rgw rgw enable ops log = true rgw enable usage log = true and you can now nc -U /tmp/rgw |./jq --stream 'fromstream(1|truncate_stream(inputs))' { "time":

[ceph-users] librados problem

2018-03-12 Thread Lizbeth Vizuet
Hello, I was trying to connect to ceph with librados and i found this problem: (0x7f4a700111d0 sd=16 :0 s=1 pgs=0 cs=0 l=1 c=0x7f4a700130b0).fault has someone have this problem? and someone knows how to solve it ___ ceph-users mailing list

Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock

2018-03-12 Thread Ilya Dryomov
On Mon, Mar 12, 2018 at 7:41 PM, Maged Mokhtar wrote: > On 2018-03-12 14:23, David Disseldorp wrote: > > On Fri, 09 Mar 2018 11:23:02 +0200, Maged Mokhtar wrote: > > 2)I undertand that before switching the path, the initiator will send a > TMF ABORT can we pass this to down

[ceph-users] bucket-notifications for ceph-rgw

2018-03-12 Thread Alex Sainer
Hi all, how do you handle bucket-notifications? is there any known piece of software we can put in front of ceph-rgw? or anything else? best regards, Alex ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock

2018-03-12 Thread Maged Mokhtar
On 2018-03-12 21:00, Ilya Dryomov wrote: > On Mon, Mar 12, 2018 at 7:41 PM, Maged Mokhtar wrote: > >> On 2018-03-12 14:23, David Disseldorp wrote: >> >> On Fri, 09 Mar 2018 11:23:02 +0200, Maged Mokhtar wrote: >> >> 2)I undertand that before switching the path, the