[ceph-users] Upgrade from Luminous to Nautilus now one MDS with could not get service secret

2021-03-28 Thread Robert LeBlanc
ything has been good until this. No other services seem to be suffering from this and it even appears that the MDS works okay even with these messages. We would like to figure out how to resolve this. Thank you, Robert LeBlanc -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904

[ceph-users] Nautilus 14.2.19 mon 100% CPU

2021-04-08 Thread Robert LeBlanc
s helped a couple of weeks ago when `ceph -s` would cause 100% CPU and lock up the service much worse than this) and didn't see the high CPU load. So I'm guessing it's triggered from some external source. I'm happy to provide more info, just let me know what would be helpful. Than

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-08 Thread Robert LeBlanc
On Thu, Apr 8, 2021 at 10:22 AM Robert LeBlanc wrote: > > I upgraded our Luminous cluster to Nautilus a couple of weeks ago and > converted the last batch of FileStore OSDs to BlueStore about 36 hours ago. > Yesterday our monitor cluster went nuts and started constantly calling

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-08 Thread Robert LeBlanc
On Thu, Apr 8, 2021 at 11:24 AM Robert LeBlanc wrote: > > On Thu, Apr 8, 2021 at 10:22 AM Robert LeBlanc wrote: > > > > I upgraded our Luminous cluster to Nautilus a couple of weeks ago and > > converted the last batch of FileStore OSDs to BlueStore about 36 hours

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-08 Thread Robert LeBlanc
mon_lease = 30 mon_osd_cache_size = 20 mon_sync_max_payload_size = 4096 -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu, Apr 8, 2021 at 1:11 PM Stefan Kooman wrote: > > On 4/8/21 6:22 PM, Robert LeBlanc wrote: > > I upgraded our Lum

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-08 Thread Robert LeBlanc
Good thought. The storage for the monitor data is a RAID-0 over three NVMe devices. Watching iostat, they are completely idle, maybe 0.8% to 1.4% for a second every minute or so. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu, Apr 8, 2021

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
I'm attempting to deep scrub all the PGs to see if that helps clear up some accounting issues, but that's going to take a really long time on 2PB of data. -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu, Apr 8, 2021 at 9:48

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
to fail over to it so I took it back out. I have run `ceph osd-require-osd-release nautilus` after the upgrade of all the OSDs. I'll go back and double check all the steps, but I think I got them all. Thank you, Robert LeBlanc ___ ceph-users mail

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
The only step not yet taken was to move to straw2. That was the last step we were going to do next. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Fri, Apr 9, 2021 at 10:41 AM Robert LeBlanc wrote: > > On Fri, Apr 9, 2021 at 9:25 AM

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
dex.php/s/OtHsBAYN9r5eSbU You can look near the end of the log for the verbose logging. I'm not sure what to look for in there, nothing sticks out to me. I did disable cephx in the config file to see if that would help, but we still have the 100% C

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
ught may help get `ceph df` working properly so we know how much storage is actually being used. We do have crons running to get metrics into graphite and some of them do ceph commands, others are creating/writing/reading/deleting files to get some file system performanc

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
On Fri, Apr 9, 2021 at 2:04 PM Dan van der Ster wrote: > > On Fri, Apr 9, 2021 at 9:37 PM Dan van der Ster wrote: > > > > On Fri, Apr 9, 2021 at 8:39 PM Robert LeBlanc wrote: > > > > > > On Fri, Apr 9, 2021 at 11:49 AM Dan van der Ster > > >

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-09 Thread Robert LeBlanc
elections. Thank you for helping us track down the bad clients out of over 2,000 clients. > > Maybe if that code path isn't needed in Nautilus it can be removed in > > the next point release? > > I think there were other major changes in this area that might make > such a back

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-11 Thread Robert LeBlanc
past dead), but a good chunk of machines are Ubuntu 14.04 not by choice. Getting this upgrade done was sorely needed, but very risky at the same time. Thanks, Robert LeBlanc ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an

[ceph-users] Re: Nautilus 14.2.19 mon 100% CPU

2021-04-12 Thread Robert LeBlanc
hink the toolchain has evolved too much since then? Thanks, Robert LeBlanc ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] 17 OSDs down simultaneously from past_interval assert

2022-03-15 Thread Robert LeBlanc
"6242'1003", "last_deep_scrub_stamp": "2022-03-14 22:53:26.138458", "last_clean_scrub_stamp": "2022-03-14 22:53:26.138458", "log_size": 1003, "ondisk_log_size": 1003,

[ceph-users] Re: 17 OSDs down simultaneously from past_interval assert

2022-03-15 Thread Robert LeBlanc
again. I had exported the first PG from the up set, but didn't wind up needing it so I didn't do it on the rest. -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Tue, Mar 15, 2022 at 10:26 AM Robert LeBlanc wrote: > We had a host

[ceph-users] 4.14 kernel or greater recommendation for multiple active MDS

2020-04-28 Thread Robert LeBlanc
ERS <https://docs.ceph.com/docs/nautilus/cephfs/kernel-features/#multiple-active-metadata-servers> The feature has been supported since the Luminous release. It is recommended to use Linux kernel clients >= 4.14 when there are multiple active MDS. Thank you, Robert LeBlanc ---

[ceph-users] Re: ceph-ansible question

2020-04-28 Thread Robert LeBlanc
evice_class: 'hdd' - data: 'blk-09' data_vg: 'ceph-blk-09' db: 'db-09' db_vg: 'ceph-db-00' crush_device_class: 'hdd' - data: 'blk-10' data_vg: 'ceph-blk-10'

[ceph-users] Re: ceph-ansible question

2020-04-29 Thread Robert LeBlanc
Yes, but they are just LVs, so you can not create them or delete them easily so that it returns the space to the VG for something else. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Tue, Apr 28, 2020 at 6:55 PM Szabo, Istvan (Agoda

[ceph-users] Re: ceph-ansible question

2020-04-30 Thread Robert LeBlanc
ple to show the worst case. -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Wed, Apr 29, 2020 at 8:17 PM Szabo, Istvan (Agoda) < istvan.sz...@agoda.com> wrote: > Ok, so this is how it works with LVM. > > > > I was pla

[ceph-users] Re: 4.14 kernel or greater recommendation for multiple active MDS

2020-05-01 Thread Robert LeBlanc
Thanks guys. We are so close to the edge that we may just take that chance, usually the only reason an active client has to reconnect is because we have to bounce the MDS because it's overwhelmed. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62

[ceph-users] Re: 4.14 kernel or greater recommendation for multiple active MDS

2020-05-02 Thread Robert LeBlanc
hanks Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Fri, May 1, 2020 at 7:37 PM Robert LeBlanc wrote: > Thanks guys. We are so close to the edge that we may just take that > chance, usually the only reason an active client has to reconnect is &

[ceph-users] Re: 4.14 kernel or greater recommendation for multiple active MDS

2020-05-07 Thread Robert LeBlanc
tubborn clients release their caps? Thanks, Robert LeBlanc -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Mon, May 4, 2020 at 7:51 PM Gregory Farnum wrote: > On Sat, May 2, 2020 at 3:12 PM Robert LeBlanc > wrote: > > > > If ther

[ceph-users] Re: Cluster blacklists MDS, can't start

2020-05-07 Thread Robert LeBlanc
On Wed, May 6, 2020 at 2:45 PM Patrick Donnelly wrote: > On Wed, Mar 11, 2020 at 10:41 PM Robert LeBlanc > wrote: > > > > This is the second time this happened in a couple of weeks. The MDS locks > > up and the stand-by can't take over so the Montiors black list

[ceph-users] Re: How many MDS servers

2020-05-07 Thread Robert LeBlanc
thub.com/ceph/ceph/pull/34571 may help if "ceph daemon mds.a > dump_mempools" shows buffer_anon uses lots of memory. > We struggle with cache management as well and I just thought it was due to our really old kernel clients (I'm sure that doesn't help). I'

[ceph-users] Re: How many MDS servers

2020-05-07 Thread Robert LeBlanc
On Thu, May 7, 2020 at 9:41 AM Robert LeBlanc wrote: > On Thu, May 7, 2020 at 6:22 AM Yan, Zheng wrote: > >> On Thu, May 7, 2020 at 1:27 AM Patrick Donnelly >> wrote: >> > >> > Hello Robert, >> > >> > On Mon, Mar 9, 2020 at 7:55 PM Rober

[ceph-users] Possible bug in op path?

2020-05-20 Thread Robert LeBlanc
on. I spent a lot of time studying the issue to come up with WPQ and saw it work great when I switched this cluster from PRIO to WPQ. I've also spent countless hours studying how it's changed in Nautilus. Thank you, Robert LeBlanc Robert LeBlanc PGP Fingerprint 79A2 9CA4

[ceph-users] Re: Possible bug in op path?

2020-05-20 Thread Robert LeBlanc
We are using high and the people on the list that have also changed have not seen the improvements that I would expect. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Wed, May 20, 2020 at 1:38 AM Dan van der Ster wrote: > > Hi

[ceph-users] Re: Possible bug in op path?

2020-05-20 Thread Robert LeBlanc
Adding the right dev list. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Wed, May 20, 2020 at 12:40 AM Robert LeBlanc wrote: > > We upgraded our Jewel cluster to Nautilus a few months ago and I've noticed > that op behavi

[ceph-users] Re: cephadm and disk partitions

2020-07-28 Thread Robert LeBlanc
VMe drives for RocksDB and then an NVMe pool for CephFS metadata. The problem with some of these configurations is that the tools don't set them up well, so I did the partitioning/LVM setup with a script I wrote, then just had Ceph use the final partition/LV. Hope that helps,

[ceph-users] Re: cephadm and disk partitions

2020-07-29 Thread Robert LeBlanc
there. The LVM layer generally is pretty light and you usually get more benefits than performance hit such relocating and extending devices online. Sorry, I can't talk to how cephadm operates. Robert LeBlanc -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3B

[ceph-users] Re: Small HDD cluster, switch from Bluestore to Filestore

2019-08-13 Thread Robert LeBlanc
st as the slowest disk), but could help with the VM with the SCSI resets. Put this in your ceph.conf and restart your OSDs. osd op queue = wpq osd op queue cut off = high Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Tue, Aug 13, 2019 at 6:

[ceph-users] Re: Small HDD cluster, switch from Bluestore to Filestore

2019-08-15 Thread Robert LeBlanc
when one drive was at a constant 100%. Did you notice any improvement in performance or system utilization in other ways? Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu, Aug 15, 2019 at 3:49 PM Rich Bade wrote: > Unfortunately the sc

[ceph-users] Re: Failing heartbeats when no backfill is running

2019-08-16 Thread Robert LeBlanc
btrfs/zfs volumes that send/receive each other. I really don't think Ceph is a good fit. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu, Aug 15, 2019 at 12:37 AM Lorenz Kiefner wrote: > Oh no, it's not that bad. It's &

[ceph-users] Re: Failing heartbeats when no backfill is running

2019-08-19 Thread Robert LeBlanc
look. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Sat, Aug 17, 2019 at 10:26 AM Lorenz Kiefner wrote: > Hello again, > > all links are at least 10/50 mbit upstream/downstream, mostly 40/100 mbit, > with some VMs at hosting compa

[ceph-users] Re: ceph fs crashes on simple fio test

2019-08-23 Thread Robert LeBlanc
The WPQ scheduler may help your clients back off when things get busy. Put this in your ceph.conf and restart your OSDs. osd op queue = wpq osd op queue cut off = high Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Fri, Aug 23, 2019 at 5:03

[ceph-users] Re: ceph fs crashes on simple fio test

2019-08-23 Thread Robert LeBlanc
We have separate public and cluster networks, but on the same physical switches with different VLANs. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Fri, Aug 23, 2019 at 1:29 PM wrote: > Frank; > > Just out of curiosity; do

[ceph-users] Re: ceph fs crashes on simple fio test

2019-08-26 Thread Robert LeBlanc
. The client latencies and backfill pain that my co-workers experienced on a daily basis have been all alleviated since moving to WPQ. Honestly, WPQ may do what you need without having to try to configure QoS. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB

[ceph-users] Re: ceph fs crashes on simple fio test

2019-08-26 Thread Robert LeBlanc
High should be the default with WPQ. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Mon, Aug 26, 2019 at 10:44 AM Paul Emmerich wrote: > WPQ has been the default queue for quite some time now (Luminous?). > > However, the defaul

[ceph-users] Re: ceph fs crashes on simple fio test

2019-08-26 Thread Robert LeBlanc
If it is the default, then the documentation should be updated. [0] [0] https://docs.ceph.com/docs/master/rados/configuration/osd-config-ref/?highlight=wpq#operations Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Mon, Aug 26, 2019 at 1:22

[ceph-users] How to test PG mapping with reweight

2019-09-02 Thread Robert LeBlanc
more distributed instead of doing a reweight, move the PGs which take a long time to just rinse and repeat over and over again. Thanks, Robert LeBlanc Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ce

[ceph-users] Re: ceph fs crashes on simple fio test

2019-09-05 Thread Robert LeBlanc
nt the client realizes that the storage is 'slow' and starts throttling the traffic and will then match the speed at which the HDDs can perform the work. If you run the job for a long time, are you still seeing the trimming errors, or just a steady rise in blocked IO on the OSDs without any

[ceph-users] Re: Best osd scenario + ansible config?

2019-09-05 Thread Robert LeBlanc
7; go down as far as 8+2 for instance and still be protected if 4 hosts in a chassis went down. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Followup: weird behaviour with ceph osd pool create and the "crush-rule" parameter (suddenly changes behaviour)

2019-09-05 Thread Robert LeBlanc
On Thu, Sep 5, 2019 at 2:12 AM wrote: > As far as I could tell, there wasn't anything in the logs (at the default > log levels), other than the forwarded request to the mons... > > (I did read the documentation, which is why I was using the long form... > it's the inconsistency which is problemat

[ceph-users] Re: How to test PG mapping with reweight

2019-09-07 Thread Robert LeBlanc
ing through all 700+ osds reweighting them. Thanks, Robert LeBlanc ---- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Mon, Sep 2, 2019 at 10:29 AM Paul Emmerich wrote: > crushtool can simulate mappings for you, it's help/man page shou

[ceph-users] Re: How to test PG mapping with reweight

2019-09-07 Thread Robert LeBlanc
ipermail/ceph-large-ceph.com/attachments/20170110/11d6382f/attachment.py [1] https://github.com/ceph/ceph/blob/cdb8df13cb7c7242f95ebbf4e404ec0006504a9f/src/osd/OSDMap.cc#L3403 [2] https://github.com/ceph/ceph/blob/f7376c0754ec700a0063bf14d0293cf31ad77798/src/osd/OSDMap.h#L768-L770 ---- Robert

[ceph-users] Re: ceph fs crashes on simple fio test

2019-09-12 Thread Robert LeBlanc
en seconds by ordinary users. > Potentially with damaging effects. I guess something similar could be > achieved with a modified rogue client. > > I would expect that a storage cluster should have basic self-defence > mechanisms that prevent this kind of overload or DOS attack by throttl

[ceph-users] RGW Passthrough

2019-09-13 Thread Robert LeBlanc
We are very interested in the RGW Passthrough mentioned for Octupus. What's the status and how can we help? We want to connect with AWS S3. Thank you, Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62

[ceph-users] Re: osds xxx have blocked requests > 1048.58 sec / osd.yyy has stuck requests > 67108.9 sec

2019-09-17 Thread Robert LeBlanc
high Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Mon, Sep 16, 2019 at 11:34 PM Thomas <74cmo...@gmail.com> wrote: > Hi, > > I have defined pool hdd which is exclusively used by virtual disks of > multiple KVMs / LXCs. > Yest

[ceph-users] RGW backup to tape

2019-09-20 Thread Robert LeBlanc
, so probably a single copy (or two copies on different tapes at the same time) when the object is created. Bonus points for being near-line. Thanks, Robert LeBlanc -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 _

[ceph-users] Re: RGW backup to tape

2019-09-20 Thread Robert LeBlanc
d all the objects that were PUT within a time frame, then read the objects off the cluster and write them to tape. Maybe that's not as easy as I'm thinking either. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1

[ceph-users] Re: cache tiering or bluestore partitions

2019-09-23 Thread Robert LeBlanc
ited for WAL (I highly recommend that for any HDD install) for each OSD, then carve out the rest as an SSD pool that you can put the CephFS metadata pool on. I don't think you would have a good experience with cache tier at that size. However, you know your access patterns far bette

[ceph-users] Re: How to reduce or control memory usage during recovery?

2019-09-23 Thread Robert LeBlanc
On Fri, Sep 20, 2019 at 5:41 AM Amudhan P wrote: > I have already set "mon osd memory target to 1Gb" and I have set max-backfill > from 1 to 8. Reducing the number of backfills should reduce the amount of memory, especially for EC pools. -------- Robert LeBlanc PGP

[ceph-users] Re: cache tiering or bluestore partitions

2019-09-23 Thread Robert LeBlanc
On Mon, Sep 23, 2019 at 10:51 AM Shawn A Kwang wrote: > > On 9/23/19 9:38 AM, Robert LeBlanc wrote: > > On Wed, Sep 18, 2019 at 11:47 AM Shawn A Kwang wrote: > >> > >> We are planning our ceph architecture and I have a question: > >> > >> How shou

[ceph-users] Re: How to reduce or control memory usage during recovery?

2019-09-24 Thread Robert LeBlanc
stripe size or object size would help, but they you may have issues in other areas. -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send a

[ceph-users] Re: Health error: 1 MDSs report slow metadata IOs, 1 MDSs report slow requests

2019-09-24 Thread Robert LeBlanc
. That my unblock things and allow things to start recovering. Then you need to fix the too_full messages by reweighting some of the full OSDs, deleting unneeded data or adding more OSDs. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 __

[ceph-users] Re: Objects degraded after adding disks

2019-10-01 Thread Robert LeBlanc
removed? If you are writing objects while the PGs are undersized (host/osds down), then it will have to sync those writes to the OSDs that were down. This is the number of degraded objects. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: one read/write, many read only

2019-10-01 Thread Robert LeBlanc
POSIX. If a client requests an exclusive write lock than no one else will be able to get a write lock. This does not mean that other clients will automatically receive a read-only lock, it depends on the client if it will fall back to read-only or fail if it can't get a write lock. ----

[ceph-users] OSD down when deleting CephFS files/leveldb compaction

2019-10-02 Thread Robert LeBlanc
urnal, but not sure if that will help in this situation. The other issue I'm seeing is that some IO just gets stuck when the OSDs are getting marked down and coming back through the cluster. Thanks, Robert LeBlanc -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E

[ceph-users] Re: OSD down when deleting CephFS files/leveldb compaction

2019-10-02 Thread Robert LeBlanc
On Wed, Oct 2, 2019 at 3:41 PM Paul Emmerich wrote: > > On Wed, Oct 2, 2019 at 10:56 PM Robert LeBlanc wrote: > > Is there a way to have leveldb compact more frequently or cause it to > > come up for air more frequently and respond to heartbeats and process > > some

[ceph-users] Re: ceph status reports: slow ops - this is related to long running process /usr/bin/ceph-osd

2019-10-08 Thread Robert LeBlanc
You have 4 OSDs that are near_full, and the errors seem to be pointed to pg_create, possibly from a backfill. Ceph will stop backfills to near_full osds. I'm guessing that is the cause of your blocked IO. Try reweighting the full OSDs down to move PGs off them. Robert LeBlan

[ceph-users] RGW blocking on large objects

2019-10-14 Thread Robert LeBlanc
ideas how to mitigate this? It seems that IOs are being scheduled on a thread and if they get unlucky enough to be scheduled behind a big IO, they are just stuck, in this case HAProxy could kick out the backend before the IO is returned and it has to re-request it. Thank you, Robe

[ceph-users] Re: RGW blocking on large objects

2019-10-15 Thread Robert LeBlanc
many RGW setups, we > default to 8192 and 800MB > > Sometimes "ms async op threads" and "ms async max op threads" might > help as well (we adjust them by default, but for other reasons) Thanks, I'll look into these options and see if they help. -

[ceph-users] Re: Dealing with changing EC Rules with drive classifications

2019-10-15 Thread Robert LeBlanc
the PG replicatas off the HDDs. Setting the following in /etc/ceph/ceph.conf on the OSDs and restarting the OSDs before backfilling will reduce the impact of the backfills. osd op queue = wpq osd op queue cut off = high Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB

[ceph-users] Re: Dealing with changing EC Rules with drive classifications

2019-10-16 Thread Robert LeBlanc
de it to a Luminous based one? > Only thing I can think that means is no upmap support, but you don't need upmap. As long as you have the CLASS column, you can do the device classes. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 __

[ceph-users] Please help me understand this large omap object found message.

2019-10-16 Thread Robert LeBlanc
tand what is going on here and what should be the best course of action. Thank you, Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: RGW blocking on large objects

2019-10-16 Thread Robert LeBlanc
On Tue, Oct 15, 2019 at 8:05 AM Robert LeBlanc wrote: > > On Mon, Oct 14, 2019 at 2:58 PM Paul Emmerich wrote: > > > > Could the 4 GB GET limit saturate the connection from rgw to Ceph? > > Simple to test: just rate-limit the health check GET > > I don't think

[ceph-users] Re: RGW blocking on large objects

2019-10-16 Thread Robert LeBlanc
On Wed, Oct 16, 2019 at 2:50 PM Paul Emmerich wrote: > > On Wed, Oct 16, 2019 at 11:23 PM Robert LeBlanc wrote: > > > > On Tue, Oct 15, 2019 at 8:05 AM Robert LeBlanc wrote: > > > > > > On Mon, Oct 14, 2019 at 2:58 PM Paul Emmerich > > > wr

[ceph-users] Re: RGW blocking on large objects

2019-10-17 Thread Robert LeBlanc
On Thu, Oct 17, 2019 at 2:50 AM Paul Emmerich wrote: > > On Thu, Oct 17, 2019 at 12:17 AM Robert LeBlanc wrote: > > > > On Wed, Oct 16, 2019 at 2:50 PM Paul Emmerich > > wrote: > > > > > > On Wed, Oct 16, 2019 at 11:23 PM Robert LeBlanc > > &g

[ceph-users] Re: RGW blocking on large objects

2019-10-17 Thread Robert LeBlanc
512 If I match the two, do you think it would help prevent small IO from being blocked by larger IO? I'm also happy to look into the code to suggest improvements if you can give me some quick points into the code to start will help. Robert LeBlanc PGP Fingerpri

[ceph-users] Re: RGW blocking on large objects

2019-10-17 Thread Robert LeBlanc
On Thu, Oct 17, 2019 at 11:46 AM Casey Bodley wrote: > > > On 10/17/19 12:59 PM, Robert LeBlanc wrote: > > On Thu, Oct 17, 2019 at 9:22 AM Casey Bodley wrote: > > > >> With respect to this issue, civetweb and beast should behave the same. > >> Both fronte

[ceph-users] Re: RGW blocking on large objects

2019-10-17 Thread Robert LeBlanc
ads in the thread pool than max concurrent requests so that there is a buffer of idle threads rather than IOs being scheduled behind one another. -------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-u

[ceph-users] Re: RGW blocking on large objects

2019-10-17 Thread Robert LeBlanc
pull/27898. It > extends the get_bucket_info() call to take optional_yield, and passes > one in where available, using null_yield to mark the synchronous cases > where one isn't available. I'll work to get familiar with the code base and see if I can submit some PRs to help out

[ceph-users] Weird blocked OP issue.

2019-11-01 Thread Robert LeBlanc
version 14.2.2 (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable)": 754, "ceph version 14.2.4 (75f4de193b3ea58512f204623e6c5a16e6c1e1ba) nautilus (stable)": 22 } } I restarted one of the monitors and it dropped out of the list only showing 2 blocked ops, but then showed up again a littl

[ceph-users] Re: Weird blocked OP issue.

2019-11-01 Thread Robert LeBlanc
On Fri, Nov 1, 2019 at 6:10 PM Robert LeBlanc wrote: > > We had an OSD host with 13 OSDs fail today and we have a weird blocked > OP message that I can't understand. There are no OSDs with blocked > ops, just `mon` (multiple times), and some of the rgw instances. >

[ceph-users] Can min_read_recency_for_promote be -1

2019-12-02 Thread Robert LeBlanc
bout promoting when there are X hits in Y time? Thanks Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Behavior of EC pool when a host goes offline

2019-12-03 Thread Robert LeBlanc
t will cause peering again. All these peerings could cause other OSDs to miss their heartbeats and the problem only gets worse as it compounds. Server failure events should be fairly infrequent that 100 seconds is a good compromise. You are able to adjust the timeout, but I highly recommend you don&

[ceph-users] Re: Can min_read_recency_for_promote be -1

2019-12-03 Thread Robert LeBlanc
t work, we have time to test it and remove the caching if it doesn't work. Thank you for your insights and I'm interested to hear what you think given the additional info. Robert LeBlanc Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Crushmap format in nautilus: documentation out of date

2019-12-05 Thread Robert LeBlanc
The second id is to overlay device types in the crush map. In the olden days, we would define two seperate roots, one for hdd and another for SSD (each host would have a duplicate entry with a slightly different name and different id), then have a crush rule use the different root for the type of s

[ceph-users] Re: RGW listing millions of objects takes too much time

2019-12-09 Thread Robert LeBlanc
actually gotten to the implementation phase of this idea. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Migrate Jewel from leveldb to rocksdb

2020-01-22 Thread Robert LeBlanc
rsion. Is there a way to move OSDs from leveldb to rocksdb without trashing and rebuilding all the osds? [0] https://docs.ceph.com/docs/master/releases/jewel/ Thanks, Robert LeBlanc Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62

[ceph-users] Re: Migrate Jewel from leveldb to rocksdb

2020-01-22 Thread Robert LeBlanc
let that run for a few days on one host before converting the others. I really hope this helps with the heavy write/delete load kicking out OSDs. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-

[ceph-users] Radosgw dynamic sharding jewel -> luminous

2020-03-02 Thread Robert LeBlanc
dmin reshard list 2020-03-02 05:31:40.875642 7fdb983f7e40 0 failed reading realm info: ret -1 (1) Operation not permitted Any nudge in the right direction would be helpful. I manually sharded the indexes, but I'd really like to have it done automatically from now on so I don't have

[ceph-users] Re: Radosgw dynamic sharding jewel -> luminous

2020-03-04 Thread Robert LeBlanc
"data_pool": "default.rgw.buckets.data", "data_extra_pool": "default.rgw.buckets.non-ec", "index_type": 0, "compression": "" } } ], "metadata_heap": "", "tier_config": [], "realm_id": "" } ``` Also the command would not work without the `--rgw-zone default` option, otherwise I'd get: ``` $ radosgw-admin zone get unable to initialize zone: (2) No such file or directory ``` Thank you, Robert LeBlanc Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Radosgw dynamic sharding jewel -> luminous

2020-03-10 Thread Robert LeBlanc
I don't know if it is that specifically, but they are all are running the latest version of Luminous and I set the cluster to only allow Luminous OSDs in. All services have been upgraded to Luminous. Do I need to run a command to activate the cls_rgw API? Robert LeBlan

[ceph-users] Cluster blacklists MDS, can't start

2020-03-11 Thread Robert LeBlanc
a pg dump, it looks like the epoch is passed that. $ ceph pg map 3.756 osdmap e234953 pg 3.756 (3.756) -> up [113,180,115] acting [113,180,115] Last time, it seemed to just recover after about an hour all by it's self. Any way to speed this up? Thank you, Robert LeBlanc --

[ceph-users] Re: Cluster blacklists MDS, can't start

2020-03-12 Thread Robert LeBlanc
the MDS got way too behind on segments ~14,000 from some naughty clients and caused the journal to explode and the MDS to eventually just not respond to the monitors. Thank you, Robert LeBlanc Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu

[ceph-users] How to get num ops blocked per OSD

2020-03-13 Thread Robert LeBlanc
d\.(\d+)$' Thanks, Robert LeBlanc ---- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: How to get num ops blocked per OSD

2020-03-14 Thread Robert LeBlanc
We are already gathering the Ceph admin socket stats with the Diamond plugin and sending that to graphite, so I guess I just need to look through that to find what I'm looking for. Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Fri, M