Re: [ceph-users] Blacklisting during boot storm

2019-08-04 Thread Kees Meijs
Hi Paul, Okay, thanks for clarifying. If we see the phenomenon again, we'll just leave it be. K. On 03-08-2019 14:33, Paul Emmerich wrote: > The usual reason for blacklisting RBD clients is breaking an exclusive > lock because the previous owner seemed to have crashed. > Blacklisting the old

[ceph-users] Blacklisting during boot storm

2019-08-03 Thread Kees Meijs
Hi list, Yesterday afternoon we experienced a compute node outage in our OpenStack (obviously Ceph backed) cluster. We tried to (re)start compute instances again as fast as possible, resulting in some KVM/RBD clients getting blacklisted. The problem was spotted very quickly so we could remove

Re: [ceph-users] Random slow requests without any load

2019-07-17 Thread Kees Meijs
Hi, Experienced similar issues. Our cluster internal network (completely separated) now has NOTRACK (no connection state tracking) iptables rules. In full: > # iptables-save > # Generated by xtables-save v1.8.2 on Wed Jul 17 14:57:38 2019 > *filter > :FORWARD DROP [0:0] > :OUTPUT ACCEPT [0:0] >

Re: [ceph-users] Altering crush-failure-domain

2019-03-04 Thread Kees Meijs
Thanks guys. Regards, Kees On 04-03-19 22:18, Smith, Eric wrote: > This will cause data migration. > > -Original Message- > From: ceph-users On Behalf Of Paul > Emmerich > Sent: Monday, March 4, 2019 2:32 PM > To: Kees Meijs > Cc: Ceph Users > Subject: Re:

[ceph-users] Altering crush-failure-domain

2019-03-04 Thread Kees Meijs
Hi Cephers, Documentation on http://docs.ceph.com/docs/master/rados/operations/erasure-code/ states: > Choosing the right profile is important because it cannot be modified > after the pool is created: a new pool with a different profile needs > to be created and all objects from the previous

Re: [ceph-users] Huge latency spikes

2018-11-17 Thread Kees Meijs
Hi Alex, What kind of clients do you use? Is it KVM (QEMU) using NBD driver, kernel, or...? Regards, Kees On 17-11-18 20:17, Alex Litvak wrote: > Hello everyone, > > I am trying to troubleshoot cluster exhibiting huge spikes of latency. > I cannot quite catch it because it happens during the

Re: [ceph-users] Ensure Hammer client compatibility

2018-11-12 Thread Kees Meijs
. There's a lot of data shuffling going on now, so fingers crossed. Cheers, Kees On 12-11-18 09:14, Kees Meijs wrote: > However, what about sortbitwise and tunables? ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/list

Re: [ceph-users] Ensure Hammer client compatibility

2018-11-12 Thread Kees Meijs
require_jewel_osds seemed harmless in terms of client compatibility so that's done already. However, what about sortbitwise and tunables? Thanks, Kees On 21-08-18 03:47, Kees Meijs wrote: We're looking at (now existing) RBD support using KVM/QEMU, so this is an upgrade path

Re: [ceph-users] Upgrade to Infernalis: OSDs crash all the time

2018-11-11 Thread Kees Meijs
and we arrived at HEALTH_OK again! Case closed: up to Jewel. For everyone involved: a big, big and even bigger thank you for all pointers and support! Regards, Kees On 10-09-18 16:43, Kees Meijs wrote: A little update: meanwhile we added a new node consisting of Hammer OSDs to ensure sufficient

Re: [ceph-users] Upgrade to Infernalis: OSDs crash all the time

2018-09-10 Thread Kees Meijs
Hi list, A little update: meanwhile we added a new node consisting of Hammer OSDs to ensure sufficient cluster capacity. The upgraded node with Infernalis OSDs is completely removed from the CRUSH map and the OSDs removed (obviously we didn't wipe the disks yet). At the moment we're still

Re: [ceph-users] Upgrade to Infernalis: OSDs crash all the time

2018-08-21 Thread Kees Meijs
Hello David, Thank you and I'm terribly sorry; I was unaware I was starting new threads. From the top of my mind I say "yes it'll fit" but obviously I make sure at first. Regards, Kees On 21-08-18 16:34, David Turner wrote: Ceph does not support downgrading OSDs.  When you removed the

Re: [ceph-users] Ensure Hammer client compatibility

2018-08-20 Thread Kees Meijs
Hi Lincoln, We're looking at (now existing) RBD support using KVM/QEMU, so this is an upgrade path. Regards, Kees On 20-08-18 16:37, Lincoln Bryant wrote: What interfaces do your Hammer clients need? If you're looking at CephFS, we have had reasonable success moving our older clients (EL6)

Re: [ceph-users] Upgrade to Infernalis: OSDs crash all the time

2018-08-20 Thread Kees Meijs
catch some sleep. Thanks, thanks! Best regards, Kees On 20-08-18 21:46, Kees Meijs wrote: Other than restarting the "out" and stopped OSD for the time being (haven't tried that yet) I'm quite lost. ___ ceph-users mailing list ceph-users

Re: [ceph-users] Upgrade to Infernalis: OSDs crash all the time

2018-08-20 Thread Kees Meijs
ain one inconsistent PG. * Backfilling started. * After hours and hours of backfilling, OSDs started to crash. Other than restarting the "out" and stopped OSD for the time being (haven't tried that yet) I'm quite lost. Hopefully someone has some pointers for me. Regards, Kees On

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-20 Thread Kees Meijs
7f8962b2f700 > time 2018-08-20 13:06:33.709922 > osd/ReplicatedPG.cc: 10115: FAILED assert(r >= 0) Restarting the OSDs seems to work. K. On 20-08-18 13:14, Kees Meijs wrote: > Bad news: I've got a PG stuck in down+peering now. ___ cep

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-20 Thread Kees Meijs
Bad news: I've got a PG stuck in down+peering now. Please advice. K. On 20-08-18 12:12, Kees Meijs wrote: > Thanks for your advice. My end goal is BlueStore so to upgrade to Jewel > and then Luminous would be ideal. > > Currently all monitors are (succesfully) running Internalis, on

[ceph-users] Ensure Hammer client compatibility

2018-08-20 Thread Kees Meijs
Good afternoon Cephers, While I'm fixing our upgrade-semi-broken cluster (see thread Upgrade to Infernalis: failed to pick suitable auth object) I'm wondering about ensuring client compatibility. My end goal is BlueStore (i.e. running Luminous) and unfortunately I'm obliged to offer Hammer

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-20 Thread Kees Meijs
Hi David, Thanks for your advice. My end goal is BlueStore so to upgrade to Jewel and then Luminous would be ideal. Currently all monitors are (succesfully) running Internalis, one OSD node is running Infernalis and all other OSD nodes have Hammer. I'll try freeing up one Infernalis OSD at

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-20 Thread Kees Meijs
Ehrm, that should of course be rebuilding. (I.e. removing the OSD, reformat, re-add.) On 20-08-18 11:51, Kees Meijs wrote: > Since there's temp in the name and we're running a 3-replica cluster, > I'm thinking of just reboiling the comprise

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-20 Thread Kees Meijs
e for one or maybe two OSDs but definitely not all. Regards, Kees On 19-08-18 08:55, Kees Meijs wrote: > I'll investigate further. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-19 Thread Kees Meijs
700 -1 log_channel(cluster) log [ERR] : 3.7d soid -5/007d/temp_3.7d_0_16204674_4402/head: failed to pick suitable auth object 2018-08-18 18:29:23.159691 7efc52942700 -1 log_channel(cluster) log [ERR] : 3.7d repair 1 errors, 0 fixed I'll investigate further. Regards, Kees On 18-08-18 17:43

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-18 Thread Kees Meijs
to fixing the ownerships while maybe still running), maybe two but not all of them. Although it's very likely it wouldn't make a difference, I'll try a ceph pg repair for each PG. To be continued again! Regards, Kees On 18-08-18 10:52, Kees Meijs wrote: To be continued... Over night, some

Re: [ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-18 Thread Kees Meijs
Hi David, Thank you for pointing out the option. On http://docs.ceph.com/docs/infernalis/release-notes/ one can read: * Ceph daemons now run as user and group ceph by default. The ceph user has a static UID assigned by Fedora and Debian (also used by derivative distributions like

[ceph-users] Upgrade to Infernalis: failed to pick suitable auth object

2018-08-17 Thread Kees Meijs
Hi Cephers, For the last months (well... years actually) we were quite happy using Hammer. So far, there was no immediate cause implying an upgrade. However, having seen Luminous providing support for BlueStore, it seemed like a good idea to perform some upgrade steps. Doing baby steps, I

Re: [ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Kees Meijs
Hi Jake, On 19-09-17 15:14, Jake Young wrote: > Ideally you actually want fewer disks per server and more servers. > This has been covered extensively in this mailing list. Rule of thumb > is that each server should have 10% or less of the capacity of your > cluster. That's very true, but let's

[ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Kees Meijs
Hi list, It's probably something to discuss over coffee in Ede tomorrow but I'll ask anyway: what HBA is best suitable for Ceph nowadays? In an earlier thread I read some comments about some "dumb" HBAs running in IT mode but still being able to use cache on the HBA. Does it make sense? Or, is

Re: [ceph-users] Shrink cache target_max_bytes

2017-02-14 Thread Kees Meijs
Hi Cephers, Although I might be stating an obvious fact: altering the parameter works as advertised. The only issue I encountered was lowering the parameter too much at once results in some slow requests because the cache pool is "full". So in short: it works when lowering the parameter bit by

[ceph-users] Shrink cache target_max_bytes

2017-02-09 Thread Kees Meijs
Hi Cephers, Long story short: I'd like to shrink our cache pool a little. Is it safe to just alter cache target_max_byte and wait for objects to get evicted? Anything to take into account? Thanks! Regards, Kees ___ ceph-users mailing list

Re: [ceph-users] All SSD cluster performance

2017-01-16 Thread Kees Meijs
Hi Maxime, Given your remark below, what kind of SATA SSD do you recommend for OSD usage? Thanks! Regards, Kees On 15-01-17 21:33, Maxime Guyot wrote: > I don’t have firsthand experience with the S3520, as Christian pointed out > their endurance doesn’t make them suitable for OSDs in most

Re: [ceph-users] Unbalanced OSD's

2016-12-30 Thread Kees Meijs
Thanks, I'll try a manual reweight at first. Have a happy new year's eve (yes, I know it's a day early)! Regards, Kees On 30-12-16 11:17, Wido den Hollander wrote: > For this reason you can do a OSD reweight by running the 'ceph osd > reweight-by-utilization' command or do it manually with

Re: [ceph-users] Unbalanced OSD's

2016-12-30 Thread Kees Meijs
Hi Asley, We experience (using Hammer) a similar issue. Not that I have a perfect solution to share, but I felt like mentioning a "me too". ;-) On a side note: we configured correct weight per drive as well. Regards, Kees On 29-12-16 11:54, Ashley Merrick wrote: > > Hello, > > > > I

Re: [ceph-users] Upgrading from Hammer

2016-12-20 Thread Kees Meijs
Hi Wido, Thanks again! Good to hear, it saves us a lot of upgrade trouble in advance. If I'm not mistaken, we haven't done anything with CRUSH tunables. Any pointers on how to make sure we really didn't? Regards, Kees On 20-12-16 10:14, Wido den Hollander wrote: > No, you don't. A Hammer/Jewel

Re: [ceph-users] Upgrading from Hammer

2016-12-20 Thread Kees Meijs
Hi Wido, At the moment, we're running Ubuntu 14.04 LTS using the Ubuntu Cloud Archive. To be precise again, it's QEMU/KVM 2.3+dfsg-5ubuntu9.4~cloud2 linked to Ceph 0.94.8-0ubuntu0.15.10.1~cloud0. So yes, it's all about running a newer QEMU/KVM on a not so new version of Ubuntu. Question is, are

[ceph-users] Upgrading from Hammer

2016-12-13 Thread Kees Meijs
Hi guys, In the past few months, I've read some posts about upgrading from Hammer. Maybe I've missed something, but I didn't really read something on QEMU/KVM behaviour in this context. At the moment, we're using: > $ qemu-system-x86_64 --version > QEMU emulator version 2.3.0 (Debian

Re: [ceph-users] 2x replication: A BIG warning

2016-12-09 Thread Kees Meijs
Hi Wido, Since it's a Friday night, I decided to just go for it. ;-) It took a while to rebalance the cache tier but all went well. Thanks again for your valuable advice! Best regards, enjoy your weekend, Kees On 07-12-16 14:58, Wido den Hollander wrote: >> Anyway, any things to consider or

Re: [ceph-users] 2x replication: A BIG warning

2016-12-07 Thread Kees Meijs
Hi Wido, Valid point. At this moment, we're using a cache pool with size = 2 and would like to "upgrade" to size = 3. Again, you're absolutely right... ;-) Anyway, any things to consider or could we just: 1. Run "ceph osd pool set cache size 3". 2. Wait for rebalancing to complete. 3. Run

[ceph-users] CoW clone performance

2016-11-25 Thread Kees Meijs
Hi list, We're using CoW clones (using OpenStack via Glance and Cinder) to store virtual machine images. For example: > # rbd info cinder-volumes/volume-a09bd74b-f100-4043-a422-5e6be20d26b2 > rbd image 'volume-a09bd74b-f100-4043-a422-5e6be20d26b2': > size 25600 MB in 3200 objects >

Re: [ceph-users] Stalling IO with cache tier

2016-11-24 Thread Kees Meijs
allow rwx pool=cinder-vms, allow rx > pool=glance-images" I presume I should add *allow rwx pool=cache* in our case? Thanks again, Kees On 24-11-16 15:55, Kees Meijs wrote: > Oh... In retrospect it makes sense in a way, but it does not as well. ;-) > > To clarify: it mak

Re: [ceph-users] Stalling IO with cache tier

2016-11-24 Thread Kees Meijs
Hi Nick, Oh... In retrospect it makes sense in a way, but it does not as well. ;-) To clarify: it makes sense since the cache is "just a pool" but it does not since "it is an overlay and just a cache in between". Anyway, something that should be well documented and warned for, if you ask me.

Re: [ceph-users] Stalling IO with cache tier

2016-11-24 Thread Kees Meijs
Hi Nick, All Ceph pools have very restrictive permissions for each OpenStack service, indeed. Besides creating the cache pool and enabling it, no additional parameters or configuration was done. Do I understand correctly access parameters (e.g. authx keys) are needed for a cache tier? If yes, it

Re: [ceph-users] Stalling IO with cache tier

2016-11-24 Thread Kees Meijs
Hi Burkhard, A testing pool makes absolute sense, thank you. About the complete setup, the documentation states: > The cache tiering agent can flush or evict objects based upon the > total number of bytes *or* the total number of objects. To specify a > maximum number of bytes, execute the

Re: [ceph-users] Stalling IO with cache tier

2016-11-24 Thread Kees Meijs
] > 13: (()+0xfbaa1) [0x5650637c1aa1] > NOTE: a copy of the executable, or `objdump -rdS ` is > needed to interpret this. > terminate called after throwing an instance of 'ceph::FailedAssertion' Hope it helps. Cheers, Kees On 24-11-16 13:06, Kees Meijs wrote: > Could someone plea

[ceph-users] Stalling IO with cache tier

2016-11-24 Thread Kees Meijs
Hi list, Our current Ceph production cluster seems to cope with performance issues, so we decided to add a fully flash based cache tier (now running with spinners and journals on separate SSDs). We ordered SSDs (Intel), disk trays and read

Re: [ceph-users] Deep scrubbing causes severe I/O stalling

2016-11-08 Thread Kees Meijs
there) a little higher. Cheers, Kees On 28-10-16 15:37, Kees Meijs wrote: > > Interesting... We're now running using deadline. In other posts I read > about noop for SSDs instead of CFQ. > > Since we're using spinners with SSD journals; does it make since to > mix the s

Re: [ceph-users] Deep scrubbing causes severe I/O stalling

2016-10-28 Thread Kees Meijs
Hi, Interesting... We're now running using deadline. In other posts I read about noop for SSDs instead of CFQ. Since we're using spinners with SSD journals; does it make since to mix the scheduler? E.g. CFG for spinners _and_ noop for SSD? K. On 28-10-16 14:43, Wido den Hollander wrote: > Make

Re: [ceph-users] Deep scrubbing causes severe I/O stalling

2016-10-28 Thread Kees Meijs
Hi, On 28-10-16 12:06, w...@42on.com wrote: > I don't like this personally. Your cluster should be capable of doing > a deep scrub at any moment. If not it will also not be able to handle > a node failure during peak times. Valid point and I totally agree. Unfortunately, the current load doesn't

[ceph-users] Deep scrubbing causes severe I/O stalling

2016-10-28 Thread Kees Meijs
Hi Cephers, Using Ceph 0.94.9-1trusty we noticed severe I/O stalling during deep scrubbing (vanilla parameters used in regards to scrubbing). I'm aware this has been discussed before, but I'd like to share the parameters we're going to evaluate: * osd_scrub_begin_hour 1 * osd_scrub_end_hour

Re: [ceph-users] Physical maintainance

2016-07-18 Thread Kees Meijs
Hi, Thanks guys, this worked like a charm. Activating the OSDs wasn't necessary: it seemed udev(7) helped me with that. Cheers, Kees On 13-07-16 14:47, Kees Meijs wrote: > So to sum up, I'd best: > > * set the noout flag > * stop the OSDs one by one > * shut down th

Re: [ceph-users] Physical maintainance

2016-07-13 Thread Kees Meijs
Thanks! So to sum up, I'd best: * set the noout flag * stop the OSDs one by one * shut down the physical node * jank the OSD drives to prevent ceph-disk(8) from automaticly activating at boot time * do my maintainance * start the physical node * reseat and activate the OSD

Re: [ceph-users] SSD Journal

2016-07-13 Thread Kees Meijs
Hi, This is an OSD box running Hammer on Ubuntu 14.04 LTS with additional systems administration tools: > $ df -h | grep -v /var/lib/ceph/osd > Filesystem Size Used Avail Use% Mounted on > udev5,9G 4,0K 5,9G 1% /dev > tmpfs 1,2G 892K 1,2G 1% /run > /dev/dm-1

Re: [ceph-users] Fwd: Re: (no subject)

2016-07-13 Thread Kees Meijs
Hi, If the qemu-img is able to handle RBD in a clever way (and I assume it does) it is able to sparsely write the image to the Ceph pool. But, it is an assumption! Maybe someone else could shed some light on this? Or even better: read the source, the RBD handler specifically. And last but not

Re: [ceph-users] Fwd: Re: (no subject)

2016-07-13 Thread Kees Meijs
Hi Fran, Fortunately, qemu-img(1) is able to directly utilise RBD (supporting sparse block devices)! Please refer to http://docs.ceph.com/docs/hammer/rbd/qemu-rbd/ for examples. Cheers, Kees On 13-07-16 09:18, Fran Barrera wrote: > Can you explain how you do this procedure? I have the same

[ceph-users] Fwd: Re: (no subject)

2016-07-13 Thread Kees Meijs
Sorry, should have posted this to the list. Forwarded Message Subject:Re: [ceph-users] (no subject) Date: Tue, 12 Jul 2016 08:30:49 +0200 From: Kees Meijs <k...@nefos.nl> To: Gaurav Goyal <er.gauravgo...@gmail.com> Hi Gaurav, It might seem

Re: [ceph-users] (no subject)

2016-07-11 Thread Kees Meijs
Glad to hear it works now! Good luck with your setup. Regards, Kees On 11-07-16 17:29, Gaurav Goyal wrote: > Hello it worked for me after removing the following parameter from > /etc/nova/nova.conf file ___ ceph-users mailing list

Re: [ceph-users] (no subject)

2016-07-11 Thread Kees Meijs
Hi, I think there's still something misconfigured: > Invalid: 400 Bad Request: Unknown scheme 'file' found in URI (HTTP 400) It seems the RBD backend is not used as expected. Have you configured both Cinder _and_ Glance to use Ceph? Regards, Kees On 08-07-16 17:33, Gaurav Goyal wrote: > > I

Re: [ceph-users] (no subject)

2016-07-08 Thread Kees Meijs
Hi, I'd recommend generating an UUID and use it for all your compute nodes. This way, you can keep your configuration in libvirt constant. Regards, Kees On 08-07-16 16:15, Gaurav Goyal wrote: > > For below section, should i generate separate UUID for both compte hosts? >

Re: [ceph-users] (no subject)

2016-07-08 Thread Kees Meijs
Hi Gaurav, Have you distributed your Ceph authentication keys to your compute nodes? And, do they have the correct permissions in terms of Ceph? K. ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Ceph cluster upgrade

2016-07-08 Thread Kees Meijs
Thank you everyone, I just tested and verified the ruleset and applied it so some pools. Worked like a charm! K. On 06-07-16 19:20, Bob R wrote: > See http://dachary.org/?p=3189 for some simple instructions on testing > your crush rule logic. ___

Re: [ceph-users] (no subject)

2016-07-08 Thread Kees Meijs
Hi Gaurav, The following snippets should suffice (for Cinder, at least): > [DEFAULT] > enabled_backends=rbd > > [rbd] > volume_driver = cinder.volume.drivers.rbd.RBDDriver > rbd_pool = cinder-volumes > rbd_ceph_conf = /etc/ceph/ceph.conf > rbd_flatten_volume_from_snapshot = false >

Re: [ceph-users] (no subject)

2016-07-07 Thread Kees Meijs
Hi Gaurav, Unfortunately I'm not completely sure about your setup, but I guess it makes sense to configure Cinder and Glance to use RBD for a backend. It seems to me, you're trying to store VM images directly on an OSD filesystem. Please refer to

Re: [ceph-users] Ceph cluster upgrade

2016-07-06 Thread Kees Meijs
Thank you very much, I'll start testing the logic prior to implementation. K. On 06-07-16 19:20, Bob R wrote: > See http://dachary.org/?p=3189 for some simple instructions on testing > your crush rule logic. ___ ceph-users mailing list

Re: [ceph-users] Ceph cluster upgrade

2016-07-06 Thread Kees Meijs
Hi Micha, Thank you very much for your prompt response. In an earlier process, I already ran: > $ ceph tell osd.* injectargs '--osd-max-backfills 1' > $ ceph tell osd.* injectargs '--osd-recovery-op-priority 1' > $ ceph tell osd.* injectargs '--osd-client-op-priority 63' > $ ceph tell osd.*

[ceph-users] Ceph cluster upgrade

2016-07-06 Thread Kees Meijs
Hi list, Given a single node Ceph cluster (lab), I started out with the following CRUSH rule: > # rules > rule replicated_ruleset { > ruleset 0 > type replicated > min_size 1 > max_size 10 > step take default > step choose firstn 0 type osd > step emit > } Meanwhile,