[ceph-users] PG numbers don't add up?

2018-03-13 Thread Nathan Dehnel
I try to add a data pool: OSD_STAT USED AVAIL TOTAL HB_PEERSPG_SUM PRIMARY_PG_SUM 9 1076M 930G 931G [0,1,2,3,4,5,6,7,8]128 5 8 1076M 930G 931G [0,1,2,3,4,5,6,7,9]128 14 7 1076M 930G 931G [0,1,2,3,4,5,6,8,9]128

Re: [ceph-users] Understanding/correcting sudden onslaught of unfound objects

2018-03-13 Thread Graham Allan
Updated cluster now to 12.2.4 and the cycle of inconsistent->repair->unfound seems to continue, though possibly slightly differently. A pg does pass through an "active+clean" phase after repair, which might be new, but more likely I never observed it at the right time before. I see messages

Re: [ceph-users] Cephfs MDS slow requests

2018-03-13 Thread Gregory Farnum
On Tue, Mar 13, 2018 at 2:56 PM David C wrote: > Thanks for the detailed response, Greg. A few follow ups inline: > > > On 13 Mar 2018 20:52, "Gregory Farnum" wrote: > > On Tue, Mar 13, 2018 at 12:17 PM, David C wrote: > >

Re: [ceph-users] Cephfs MDS slow requests

2018-03-13 Thread David C
Thanks for the detailed response, Greg. A few follow ups inline: On 13 Mar 2018 20:52, "Gregory Farnum" wrote: On Tue, Mar 13, 2018 at 12:17 PM, David C wrote: > Hi All > > I have a Samba server that is exporting directories from a Cephfs Kernel >

Re: [ceph-users] Cephfs MDS slow requests

2018-03-13 Thread Gregory Farnum
On Tue, Mar 13, 2018 at 12:17 PM, David C wrote: > Hi All > > I have a Samba server that is exporting directories from a Cephfs Kernel > mount. Performance has been pretty good for the last year but users have > recently been complaining of short "freezes", these seem to

Re: [ceph-users] Issue with fstrim and Nova hw_disk_discard=unmap

2018-03-13 Thread Jason Dillaman
Can you provide the output from "rbd info /volume-80838a69-e544-47eb-b981-a4786be89736"? On Tue, Mar 13, 2018 at 12:30 PM, Fulvio Galeazzi wrote: > Hallo! > >> Discards appear like they are being sent to the device. How big of a >> temporary file did you create and then

Re: [ceph-users] Issue with fstrim and Nova hw_disk_discard=unmap

2018-03-13 Thread Jason Dillaman
Discards appear like they are being sent to the device. How big of a temporary file did you create and then delete? Did you sync the file to disk before deleting it? What version of qemu-kvm are you running? On Tue, Mar 13, 2018 at 11:09 AM, Fulvio Galeazzi wrote: >

Re: [ceph-users] SSD as DB/WAL performance with/without drive write cache

2018-03-13 Thread Caspar Smit
Hi Vladimir, Yeah, the results are pretty low compared to yours but i think this is due to the fact that this SSD is in a fairly old server (Supermicro X8, SAS2 expander backplane). Controller is LSI/Broadcom 9207-8i on the latest IT firmware (Same LSI2308 chipset as yours) Kind regards, Caspar

Re: [ceph-users] SSD as DB/WAL performance with/without drive write cache

2018-03-13 Thread Дробышевский , Владимир
Hello, Caspar! Would you mind to share controller model you use? I would say these results are pretty low. Here are my results on Intel RMS25LB LSI2308 based SAS controller in IT mode: I set write_cache to write through Test command, fio 2.2.10: sudo fio --filename=/dev/sdb --direct=1

Re: [ceph-users] Civetweb log format

2018-03-13 Thread Aaron Bassett
Well I have it mostly wrapped up and writing to graylog, however the ops log has a `remote_addr` field, but as far as I can tell it's always blank. I found this fix but it seems to only be in v13.0.1 https://github.com/ceph/ceph/pull/16860 Is there any chance we'd see backports of this to

[ceph-users] Cephfs MDS slow requests

2018-03-13 Thread David C
Hi All I have a Samba server that is exporting directories from a Cephfs Kernel mount. Performance has been pretty good for the last year but users have recently been complaining of short "freezes", these seem to coincide with MDS related slow requests in the monitor ceph.log such as: 2018-03-13

Re: [ceph-users] Fwd: [ceph bad performance], can't find a bottleneck

2018-03-13 Thread Sergey Kotov
Hi, Maged Not a big difference in both cases. Performance of 4 nodes pool with 5x PM863a each is: 4k bs - 33-37kIOPS krbd 128 threads and 42-51kIOPS vs 1024 threads (fio numjobs 128-256-512) the same situation happens when we try to increase rbd workload, 3 rbd gets the same iops #. Dead end &

Re: [ceph-users] Object Gateway - Server Side Encryption

2018-03-13 Thread Casey Bodley
On 03/10/2018 12:58 AM, Amardeep Singh wrote: On Saturday 10 March 2018 02:01 AM, Casey Bodley wrote: On 03/08/2018 07:16 AM, Amardeep Singh wrote: Hi, I am trying to configure server side encryption using Key Management Service as per documentation

[ceph-users] SSD as DB/WAL performance with/without drive write cache

2018-03-13 Thread Caspar Smit
Hi all, I've tested some new Samsung SM863 960GB and Intel DC S4600 240GB SSD's using the method described at Sebastien Han's blog: https://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/ The first thing stated there is to disable the drive's

Re: [ceph-users] Issue with fstrim and Nova hw_disk_discard=unmap

2018-03-13 Thread Fulvio Galeazzi
Hallo! Discards appear like they are being sent to the device. How big of a temporary file did you create and then delete? Did you sync the file to disk before deleting it? What version of qemu-kvm are you running? I made several test with commands like (issuing sync after each operation):

Re: [ceph-users] Issue with fstrim and Nova hw_disk_discard=unmap

2018-03-13 Thread Fulvio Galeazzi
Hallo Jason, thanks for your feedback! Original Message >> * decorated a CentOS image with hw_scsi_model=virtio--scsi,hw_disk_bus=scsi> > Is that just a typo for "hw_scsi_model"? Yes, it was a typo when I wrote my message. The image has virtio-scsi as it should. I

Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock

2018-03-13 Thread Ilya Dryomov
On Mon, Mar 12, 2018 at 8:20 PM, Maged Mokhtar wrote: > On 2018-03-12 21:00, Ilya Dryomov wrote: > > On Mon, Mar 12, 2018 at 7:41 PM, Maged Mokhtar wrote: > > On 2018-03-12 14:23, David Disseldorp wrote: > > On Fri, 09 Mar 2018 11:23:02 +0200, Maged

Re: [ceph-users] New Ceph cluster design

2018-03-13 Thread Christian Balzer
Hello, On Sat, 10 Mar 2018 16:14:53 +0100 Vincent Godin wrote: > Hi, > > As i understand it, you'll have one RAID1 of two SSDs for 12 HDDs. A > WAL is used for all writes on your host. This isn't filestore, AFAIK the WAL/DB will be used for small writes only to keep latency with Bluestore