Re: [ceph-users] Is it possible to recover from block.db failure?

2017-10-19 Thread Wido den Hollander
> Op 19 oktober 2017 om 16:47 schreef Caspar Smit : > > > Hi David, > > Thank you for your answer, but wouldn't scrub (deep-scrub) handle > that? It will flag the unflushed journal pg's as inconsistent and you > would have to repair the pg's. Or am i overlooking

Re: [ceph-users] High mem with Luminous/Bluestore

2017-10-18 Thread Wido den Hollander
> Op 18 oktober 2017 om 11:41 schreef Hans van den Bogert > : > > > Hi All, > > I've converted 2 nodes with 4 HDD/OSDs each from Filestore to Bluestore. I > expected somewhat higher memory usage/RSS values, however I see, imo, a > huge memory usage for all OSDs on both

Re: [ceph-users] Bluestore OSD_DATA, WAL & DB

2017-10-18 Thread Wido den Hollander
> Op 17 oktober 2017 om 14:21 schreef Mark Nelson <mnel...@redhat.com>: > > > > > On 10/17/2017 01:54 AM, Wido den Hollander wrote: > > > >> Op 16 oktober 2017 om 18:14 schreef Richard Hesketh > >> <richard.hesk...@rd.bbc.co.uk>: >

Re: [ceph-users] Thick provisioning

2017-10-18 Thread Wido den Hollander
> Op 17 oktober 2017 om 19:38 schreef Jason Dillaman : > > > There is no existing option to thick provision images within RBD. When > an image is created or cloned, the only actions that occur are some > small metadata updates to describe the image. This allows image >

Re: [ceph-users] Luminous : 3 clients failing to respond to cache pressure

2017-10-17 Thread Wido den Hollander
> Op 17 oktober 2017 om 15:35 schreef Yoann Moulin : > > > Hello, > > I have a luminous (12.2.1) cluster with 3 nodes for cephfs (no rbd or rgw) > and we hit the "X clients failing to respond to cache pressure" message. > I have 3 mds servers active. > What type of

Re: [ceph-users] How to get current min-compat-client setting

2017-10-17 Thread Wido den Hollander
unless someone thinks > it’s not a good idea. > I think it's a good idea! This probably needs to go into mon/OSDMonitor.cc The OSDMap class has a get_min_compat_client() function which you can probably use. Wido > Regards > > Hans > > > On Oct 16, 2017,

Re: [ceph-users] Bluestore OSD_DATA, WAL & DB

2017-10-17 Thread Wido den Hollander
> Op 16 oktober 2017 om 18:14 schreef Richard Hesketh > <richard.hesk...@rd.bbc.co.uk>: > > > On 16/10/17 13:45, Wido den Hollander wrote: > >> Op 26 september 2017 om 16:39 schreef Mark Nelson <mnel...@redhat.com>: > >> On 09/26/2017 01:10

Re: [ceph-users] Bluestore OSD_DATA, WAL & DB

2017-10-16 Thread Wido den Hollander
> Op 26 september 2017 om 16:39 schreef Mark Nelson : > > > > > On 09/26/2017 01:10 AM, Dietmar Rieder wrote: > > thanks David, > > > > that's confirming what I was assuming. To bad that there is no > > estimate/method to calculate the db partition size. > > It's possible

Re: [ceph-users] rados export/import fail

2017-10-16 Thread Wido den Hollander
> Op 16 oktober 2017 om 13:00 schreef Nagy Ákos : > > > Thanks, > > but I erase all of the data, I have only this backup. I hate to bring the bad news, but it will not work. The pools have different IDs and that will make it very difficult to get this working again.

Re: [ceph-users] How to get current min-compat-client setting

2017-10-16 Thread Wido den Hollander
> Op 13 oktober 2017 om 10:22 schreef Hans van den Bogert > : > > > Hi, > > I’m in the middle of debugging some incompatibilities with an upgrade of > Proxmox which uses Ceph. At this point I’d like to know what my current value > is for the min-compat-client setting,

Re: [ceph-users] Bluestore "separate" WAL and DB

2017-10-16 Thread Wido den Hollander
I thought I'd pick up on this older thread instead of starting a new one. For the WAL something between 512MB and 2GB should be sufficient as Mark Nelson explained in a different thread. The DB however I'm not certain about at this moment. The general consensus seems to be: "use as much as

Re: [ceph-users] Ceph mirrors

2017-10-06 Thread Wido den Hollander
> Op 5 oktober 2017 om 21:35 schreef Stefan Kooman : > > > Hi, > > Sorry for empty mail, that shouldn't have happened. I would like to > address the following. Currently the repository list for debian- > packages contain _only_ the latest package version. In case of a > (urgent)

Re: [ceph-users] TLS for tracker.ceph.com

2017-10-05 Thread Wido den Hollander
> Op 5 oktober 2017 om 15:57 schreef Stefan Kooman : > > > Hi, > > Can we supply http://tracker.ceph.com with TLS and make it > https://tracker.ceph.com? Should be trivial with Let's Encrypt for > example. > > Thanks! Yes please! This is something which has been asked for a

Re: [ceph-users] librmb: Mail storage on RADOS with Dovecot

2017-09-25 Thread Wido den Hollander
> Op 22 september 2017 om 23:56 schreef Gregory Farnum : > > > On Fri, Sep 22, 2017 at 2:49 PM, Danny Al-Gaaf > wrote: > > Am 22.09.2017 um 22:59 schrieb Gregory Farnum: > > [..] > >> This is super cool! Is there anything written down that explains

Re: [ceph-users] librmb: Mail storage on RADOS with Dovecot

2017-09-22 Thread Wido den Hollander
ore practical. > Great! Keep in mind this is still in a very experimental phase, but we can use all the feedback to make librmb awesome. Issues can be reported on Github. Thanks! Wido > > > -Original Message- > > From: ceph-users [mailto:ceph-users-boun...@lis

[ceph-users] librmb: Mail storage on RADOS with Dovecot

2017-09-21 Thread Wido den Hollander
Hi, A tracker issue has been out there for a while: http://tracker.ceph.com/issues/12430 Storing e-mail in RADOS with Dovecot, the IMAP/POP3/LDA server with a huge marketshare. It took a while, but last year Deutsche Telekom took on the heavy work and started a project to develop librmb:

Re: [ceph-users] What HBA to choose? To expand or not to expand?

2017-09-19 Thread Wido den Hollander
> Op 19 september 2017 om 13:34 schreef Kees Meijs : > > > Hi list, > > It's probably something to discuss over coffee in Ede tomorrow but I'll > ask anyway: what HBA is best suitable for Ceph nowadays? > I still prefer LSI (Avago) in most systems. A 8-port or 16-port

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Wido den Hollander
> Op 19 september 2017 om 10:24 schreef Adrian Saul > : > > > > I understand what you mean and it's indeed dangerous, but see: > > https://github.com/ceph/ceph/blob/master/systemd/ceph-osd%40.service > > > > Looking at the systemd docs it's difficult though: > >

Re: [ceph-users] ceph-osd restartd via systemd in case of disk error

2017-09-19 Thread Wido den Hollander
> Op 19 september 2017 om 10:02 schreef Manuel Lausch : > > > Hi, > > I see a issue with systemd's restart behaviour and disk IO-errors > If a disk fails with IO-errors ceph-osd stops running. Systemd detects > this and starts the daemon again. In our cluster I did see

Re: [ceph-users] Ceph 12.2.0 and replica count

2017-09-18 Thread Wido den Hollander
> Op 18 september 2017 om 10:14 schreef Max Krasilnikov : > > > Hello! > > In the times of Hammer it was actual to have 3 replicas for data to avoid > situation with non-identical data on different OSDs. Now we have full data and > metadata checksumming. So, is it actual

Re: [ceph-users] moving mons across networks

2017-09-13 Thread Wido den Hollander
> Op 13 september 2017 om 10:38 schreef Dan van der Ster : > > > Hi Blair, > > You can add/remove mons on the fly -- connected clients will learn > about all of the mons as the monmap changes and there won't be any > downtime as long as the quorum is maintained. > > There

Re: [ceph-users] How big can a mon store get?

2017-08-25 Thread Wido den Hollander
> Op 25 augustus 2017 om 15:00 schreef Matthew Vernon : > > > Hi, > > We have a medium-sized (2520 osds, 42 hosts, 88832 pgs, 15PB raw > capacity) Jewel cluster (on Ubuntu), and in normal operation, our mon > store size is around the 1.2G mark. I've noticed, though, that

[ceph-users] Ceph Day Netherlands: 20-09-2017

2017-08-24 Thread Wido den Hollander
Hi, In less then a month the Ceph Day in NL is coming up. It will be hosted by BIT [0] at their great venue in Ede, NL. The schedule hasn't been posted yet, we are still working on that. There will be a great talk from the people of BIT showing off their SSD-only cluster spread out over

Re: [ceph-users] luminous/bluetsore osd memory requirements

2017-08-10 Thread Wido den Hollander
> Op 10 augustus 2017 om 11:14 schreef Marcus Haarmann > : > > > Hi, > > we have done some testing with bluestore and found that the memory > consumption of the osd > processes is depending not on the real data amount stored but on the number > of stored >

Re: [ceph-users] All flash ceph witch NVMe and SPDK

2017-08-07 Thread Wido den Hollander
> Op 3 augustus 2017 om 15:28 schreef Mike A : > > > Hello > > Our goal it is make fast storage as possible. > By now our configuration of 6 servers look like that: > * 2 x CPU Intel Gold 6150 20 core 2.4Ghz > * 2 x 16 Gb NVDIMM DDR4 DIMM > * 6 x 16 Gb RAM DDR4 > * 6

Re: [ceph-users] Gracefully reboot OSD node

2017-08-03 Thread Wido den Hollander
> Op 3 augustus 2017 om 14:14 schreef Hans van den Bogert > : > > > Thanks for answering even before I asked the questions:) > > So bottom line, HEALTH_ERR state is simply part of taking a (bunch of) OSD > down? Is HEALTH_ERR period of 2-4 seconds within normal bounds?

Re: [ceph-users] Gracefully reboot OSD node

2017-08-03 Thread Wido den Hollander
> Op 3 augustus 2017 om 13:36 schreef linghucongsong : > > > > > set the osd noout nodown > While noout is correct and might help in some situations, never set nodown unless you really need that. It will block I/O since you are taking down OSDs which aren't marked

Re: [ceph-users] CEPH bluestore space consumption with small objects

2017-08-03 Thread Wido den Hollander
> Op 2 augustus 2017 om 17:55 schreef Marcus Haarmann > : > > > Hi, > we are doing some tests here with a Kraken setup using bluestore backend (on > Ubuntu 64 bit). > We are trying to store > 10 mio very small objects using RADOS. > (no fs, no rdb, only osd and

[ceph-users] Override SERVER_PORT and SERVER_PORT_SECURE and AWS4

2017-08-01 Thread Wido den Hollander
Hi, I'm running into a issue with RGW running Civetweb behind a Apache mod_proxy server. The problem is that when AWS credentials and signatures are send using the Query String the host header calculated by RGW is something like this: host:rgw.mydomain.local:7480 RGW thinks it's running on

Re: [ceph-users] ceph-mon not listening on IPv6?

2017-07-31 Thread Wido den Hollander
> Op 31 juli 2017 om 11:40 schreef Stuart Longland <stua...@longlandclan.id.au>: > > > On 31/07/17 19:10, Wido den Hollander wrote: > > > >> Op 30 juli 2017 om 2:42 schreef Stuart Longland > >> <stua...@longlandclan.id.au>: > &g

Re: [ceph-users] ceph-mon not listening on IPv6?

2017-07-31 Thread Wido den Hollander
> Op 30 juli 2017 om 2:42 schreef Stuart Longland : > > > Hi all, > > I'm setting up an experimental cluster at home, consisting of 3 nodes > which run combined ceph-osd and ceph-mon daemons, and a pair of nodes > that run virtual machines. > > All 5 nodes run

Re: [ceph-users] Defining quota in CephFS - quota is ignored

2017-07-26 Thread Wido den Hollander
> Op 26 juli 2017 om 13:43 schreef c.mo...@web.de: > > > 26. Juli 2017 11:29, "Wido den Hollander" <w...@42on.com> schrieb: > > >> Op 26 juli 2017 om 11:26 schreef c.mo...@web.de: > >> > >> Hello! > >> > >> Based on

Re: [ceph-users] Defining quota in CephFS - quota is ignored

2017-07-26 Thread Wido den Hollander
> Op 26 juli 2017 om 11:26 schreef c.mo...@web.de: > > > Hello! > > Based on the documentation for defining quotas in CephFS for any directory > (http://docs.ceph.com/docs/master/cephfs/quota/), I defined a quota for > attribute max_bytes: > ld4257:~ # getfattr -n ceph.quota.max_bytes

Re: [ceph-users] ipv6 monclient

2017-07-19 Thread Wido den Hollander
> Op 19 juli 2017 om 10:36 schreef Dan van der Ster : > > > Hi Wido, > > Quick question about IPv6 clusters which you may have already noticed. > We have an IPv6 cluster and clients use this as the ceph.conf: > > [global] > mon host = cephv6.cern.ch > > cephv6 is an

Re: [ceph-users] Modify pool size not allowed with permission osd 'allow rwx pool=test'

2017-07-18 Thread Wido den Hollander
> Op 18 juli 2017 om 17:40 schreef Marc Roos : > > > > > With ceph auth I have set permissions like below, I can add and delete > objects in the test pool, but cannot set size of a the test pool. What > permission do I need to add for this user to modify the size

Re: [ceph-users] Ceph (Luminous) shows total_space wrong

2017-07-17 Thread Wido den Hollander
seen as > 200GB. > > How can I make sure BlueStore is disabled (even if i not put any command). > Just use BlueStore with Luminous as all testing is welcome! But in this case you invoked the command with the wrong parameters. Wido > -Gencer. > > -Original Message- &g

Re: [ceph-users] Ceph (Luminous) shows total_space wrong

2017-07-17 Thread Wido den Hollander
174 > 11 0.00980 1.0 10240M 1068M 9171M 10.44 1.00 153 > 13 0.00980 1.0 10240M 1068M 9171M 10.44 1.00 179 > 15 0.00980 1.0 10240M 1068M 9171M 10.44 1.00 186 > 17 0.00980 1.0 10240M 1068M 9171M 10.44 1.00 185 > 19 0.00980 1.0 10240M 1067M 9172M 10.43 1.0

Re: [ceph-users] Ceph (Luminous) shows total_space wrong

2017-07-17 Thread Wido den Hollander
> Op 17 juli 2017 om 15:49 schreef gen...@gencgiyen.com: > > > Hi, > > > > I successfully managed to work with ceph jewel. Want to try luminous. > > > > I also set experimental bluestore while creating osds. Problem is, I have > 20x3TB hdd in two nodes and i would expect 55TB usable (as

Re: [ceph-users] autoconfigured haproxy service?

2017-07-14 Thread Wido den Hollander
> Op 11 juli 2017 om 22:35 schreef Sage Weil <sw...@redhat.com>: > > > On Tue, 11 Jul 2017, Wido den Hollander wrote: > > > Op 11 juli 2017 om 17:03 schreef Sage Weil <sw...@redhat.com>: > > > > > > > > > Hi all, > > > >

Re: [ceph-users] autoconfigured haproxy service?

2017-07-11 Thread Wido den Hollander
> Op 11 juli 2017 om 17:03 schreef Sage Weil : > > > Hi all, > > Luminous features a new 'service map' that lets rgw's (and rgw nfs > gateways and iscsi gateways and rbd mirror daemons and ...) advertise > themselves to the cluster along with some metadata (like the

Re: [ceph-users] Ceph MeetUp Berlin on July 17

2017-07-11 Thread Wido den Hollander
Hi, I would have loved to join, but it's a bit of short notice to travel from the Netherlands :-) Wido > Op 10 juli 2017 om 9:39 schreef Robert Sander : > > > Hi, > > https://www.meetup.com/de-DE/Ceph-Berlin/events/240812906/ > > Come join us for an

Re: [ceph-users] osdmap several thousand epochs behind latest

2017-07-11 Thread Wido den Hollander
> Op 10 juli 2017 om 2:06 schreef Chris Apsey : > > > All, > > Had a fairly substantial network interruption that knocked out about > ~270 osds: > > health HEALTH_ERR > [...] > 273/384 in osds are down > noup,nodown,noout

Re: [ceph-users] Watch for fstrim running on your Ubuntu systems

2017-07-08 Thread Wido den Hollander
0 -1 common/HeartbeatMap.cc > >>> <http://heartbeatmap.cc/>: In function 'bool > >>> ceph::HeartbeatMap::_check(const ceph::heartbeat_handle_d*, const char*, > >>> time_t)' thread 7ff612721700 time 2017-07-02 13:47:32.618784 > >>> common/

Re: [ceph-users] New cluster - configuration tips and reccomendation - NVMe

2017-07-06 Thread Wido den Hollander
tarts counting if you really want to push IOps. Especially over NVMe. You will need very fast CPUs to fully utilize those cards. Wido > > > > Il 05/07/2017 12:51, Wido den Hollander ha scritto: > >> Op 5 juli 2017 om 12:39 schreef c...@jack.fr.eu.org: > &

Re: [ceph-users] New cluster - configuration tips and reccomendation - NVMe

2017-07-05 Thread Wido den Hollander
ow, that increases to 25k during higher load, but the bandwidth stays below 500MB/sec in TOTAL. So yes, you are right, a NVMe device can sature a single NIC, but most of the time latency and IOps are what count. Not bandwidth. Wido > On 05/07/2017 11:54, Wido den Hollander wrote: > > &g

Re: [ceph-users] bluestore behavior on disks sector read errors

2017-07-05 Thread Wido den Hollander
> Op 27 juni 2017 om 11:17 schreef SCHAER Frederic : > > > Hi, > > Every now and then , sectors die on disks. > When this happens on my bluestore (kraken) OSDs, I get 1 PG that becomes > degraded. > The exact status is : > > > HEALTH_ERR 1 pgs inconsistent; 1 scrub

Re: [ceph-users] New cluster - configuration tips and reccomendation - NVMe

2017-07-05 Thread Wido den Hollander
> Op 5 juli 2017 om 10:54 schreef Massimiliano Cuttini : > > > Dear all, > > luminous is coming and sooner we should be allowed to avoid double writing. > This means use 100% of the speed of SSD and NVMe. > Cluster made all of SSD and NVMe will not be penalized and start to

Re: [ceph-users] Cache Tier or any other possibility to accelerate RBD with SSD?

2017-07-03 Thread Wido den Hollander
> Op 3 juli 2017 om 13:01 schreef Mateusz Skała : > > > Hello, > > We are using cache-tier in Read-forward mode (replica 3) for accelerate > reads and journals on SSD to accelerate writes. We are using only RBD. Based > on the ceph-docs, RBD have bad I/O pattern for

Re: [ceph-users] Any recommendations for CephFS metadata/data pool sizing?

2017-07-01 Thread Wido den Hollander
> Op 1 juli 2017 om 1:04 schreef Tu Holmes : > > > I would use the calculator at ceph and just set for "all in one". > > http://ceph.com/pgcalc/ > I wouldn't do that. With CephFS the data pool(s) will contain much more objects and data then the metadata pool. You can

Re: [ceph-users] dropping filestore+btrfs testing for luminous

2017-06-30 Thread Wido den Hollander
> Op 30 juni 2017 om 18:48 schreef Sage Weil : > > > On Fri, 30 Jun 2017, Lenz Grimmer wrote: > > Hi Sage, > > > > On 06/30/2017 05:21 AM, Sage Weil wrote: > > > > > The easiest thing is to > > > > > > 1/ Stop testing filestore+btrfs for luminous onward. We've recommended

Re: [ceph-users] luminous v12.1.0 bluestore by default doesnt work

2017-06-30 Thread Wido den Hollander
> Op 30 juni 2017 om 13:35 schreef Малков Петр Викторович : > > > v12.1.0 Luminous RC released > BlueStore: > The new BlueStore backend for ceph-osd is now stable and the new > default for newly created OSDs. > > [global] > fsid = a737f8ad-b959-4d44-ada7-2ed6a2b8802b >

Re: [ceph-users] Zabbix plugin for ceph-mgr

2017-06-29 Thread Wido den Hollander
Just opened a PR: https://github.com/ceph/ceph/pull/16019 Reviews and comments are welcome! Wido > Op 27 juni 2017 om 16:57 schreef Wido den Hollander <w...@42on.com>: > > > > > Op 27 juni 2017 om 16:13 schreef David Turner <drakonst...@gmail.com>: >

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-29 Thread Wido den Hollander
nd IPv6, but the Ceph cluster is IPv6-only. RGW/librados talks to Ceph ovre IPv6 and handles client traffic on both protocols. No problem to run the RGW dual-stacked. Wido > Brenno > > > Em 29/06/2017 03:13:26, Wido den Hollander escreveu: > > > Op 28

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-29 Thread Wido den Hollander
- The OSDMap/MONMap can only contain ONE IP, eg, IPv4 or IPv6 - Messengers can't talk dual-stacked So you have to shut down the whole cluster and update the MONMap. Afterwards the OSDs will boot and 'up' themselves in the cluster with their IPv6 address and that will be recorded in the OSDMap.

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-28 Thread Wido den Hollander
Best regards, > > George > > From: Wido den Hollander [w...@42on.com] > Sent: 27 June 2017 19:19 > To: Vasilakakos, George (STFC,RAL,SC); ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Ceph and IPv4 -> IPv6 > > > O

Re: [ceph-users] Upgrade target for 0.82

2017-06-27 Thread Wido den Hollander
> Op 27 juni 2017 om 20:56 schreef German Anders : > > > Hi Cephers, > >I want to upgrade an existing cluster (version 0.82), and I would like > to know if there's any recommended upgrade-path and also the recommended > target version. > I would go to Hammer (0.94)

Re: [ceph-users] Ceph and IPv4 -> IPv6

2017-06-27 Thread Wido den Hollander
> Op 27 juni 2017 om 19:00 schreef george.vasilaka...@stfc.ac.uk: > > > Hey Ceph folks, > > I was wondering what the current status/roadmap/intentions etc. are on the > possibility of providing a way of transitioning a cluster from IPv4 to IPv6 > in the future. > > My current understanding

Re: [ceph-users] Zabbix plugin for ceph-mgr

2017-06-27 Thread Wido den Hollander
There's another one based on python. It got me 90% of the way to > where I wanted Zabbix to be for monitoring Ceph. > https://github.com/thelan/ceph-zabbix > > On Tue, Jun 27, 2017 at 5:59 AM Wido den Hollander <w...@42on.com> wrote: > > > > > > Op 27 juni 2017 om 11:24 schr

Re: [ceph-users] TRIM/Discard on SSDs with BlueStore

2017-06-27 Thread Wido den Hollander
> Op 27 juni 2017 om 13:56 schreef Christian Balzer <ch...@gol.com>: > > > On Tue, 27 Jun 2017 13:24:45 +0200 (CEST) Wido den Hollander wrote: > > > > Op 27 juni 2017 om 13:05 schreef Christian Balzer <ch...@gol.com>: > > > > > > &g

Re: [ceph-users] TRIM/Discard on SSDs with BlueStore

2017-06-27 Thread Wido den Hollander
> Op 27 juni 2017 om 13:26 schreef Mark Nelson <mnel...@redhat.com>: > > > > > On 06/27/2017 06:24 AM, Wido den Hollander wrote: > > > >> Op 27 juni 2017 om 13:05 schreef Christian Balzer <ch...@gol.com>: > >> > >> > >

Re: [ceph-users] Zabbix plugin for ceph-mgr

2017-06-27 Thread Wido den Hollander
ode. Then I moved to nagios. > I see Zabbix envs running with thousands of hosts and 10ks of items in there without any issues. It's ofcourse a personal preference. Working at a location now who are eager to go to Luminous and would like to see such a module for ceph-mgr. Wido > >

[ceph-users] TRIM/Discard on SSDs with BlueStore

2017-06-27 Thread Wido den Hollander
Hi, I've been looking in the docs and the source code of BlueStore to figure out if it issues TRIM/Discard [0] on SSDs and I haven't been able to find an answer. Does BlueStore/BlueFS issue these commands to give back the space to the underlying device? For SSDs it improves both write

[ceph-users] Zabbix plugin for ceph-mgr

2017-06-27 Thread Wido den Hollander
Hi, After looking at the documentation [0] on how to write a plugin for ceph-mgr I've been playing with the idea to create a Zabbix [1] plugin for ceph-mgr. Before I start writing one I'd like to check if I'm thinking in the right direction. Zabbix supports Items [2] and Triggers. Triggers

Re: [ceph-users] Ideas on the UI/UX improvement of ceph-mgr: Cluster Status Dashboard

2017-06-27 Thread Wido den Hollander
> Op 26 juni 2017 om 6:49 schreef saumay agrawal : > > > Hi everyone! > > I am working on the improvement of the web-based dashboard for Ceph. > My intention is to add some UI elements to visualise some performance > counters of a Ceph cluster. This gives a better

[ceph-users] ceph-mon not starting on Ubuntu 16.04 with Luminous RC

2017-06-27 Thread Wido den Hollander
Hi, Just checking before I start looking into ceph-deploy if the behavior I'm seeing is correct. On a freshly installed Ubuntu 16.04 + Luminous 12.1.0 system I see that my ceph-mon services aren't starting on boot. Deployed Ceph on three machines: alpha, bravo and charlie. Using 'alpha' I've

Re: [ceph-users] v12.1.0 Luminous RC released

2017-06-25 Thread Wido den Hollander
> Op 23 juni 2017 om 23:06 schreef Sage Weil : > > > On Fri, 23 Jun 2017, Abhishek L wrote: > > This is the first release candidate for Luminous, the next long term > > stable release. > > I just want to reiterate that this is a release candidate, not the final > luminous

Re: [ceph-users] Flash for mon nodes ?

2017-06-21 Thread Wido den Hollander
> Op 21 juni 2017 om 12:38 schreef Osama Hasebou : > > > Hi Guys, > > Has anyone used flash SSD drives for nodes hosting Monitor nodes only? > > If yes, any major benefits against just using SAS drives ? > Yes: - Less latency - Faster store compacting - More

Re: [ceph-users] FAILED assert(i.first <= i.last)

2017-06-19 Thread Wido den Hollander
ed with the files on the disk and then had data corruption. That's why I'd say that you should try to avoid fixing single OSDs when you don't need to. Wido > /Peter > > mån 19 juni 2017 kl 09:11 skrev Wido den Hollander <w...@42on.com>: > > > > > > Op 18

Re: [ceph-users] Kernel RBD client talking to multiple storage clusters

2017-06-19 Thread Wido den Hollander
> Op 19 juni 2017 om 5:15 schreef Alex Gorbachev : > > > Has anyone run into such config where a single client consumes storage from > several ceph clusters, unrelated to each other (different MONs and OSDs, > and keys)? > Should be possible, you can simply supply a

Re: [ceph-users] FAILED assert(i.first <= i.last)

2017-06-19 Thread Wido den Hollander
> Op 18 juni 2017 om 16:21 schreef Peter Rosell : > > > Hi, > I have a small cluster with only three nodes, 4 OSDs + 3 OSDs. I have been > running version 0.87.2 (Giant) for over 2.5 year, but a couple of day ago I > upgraded to 0.94.10 (Hammer) and then up to 10.2.7

Re: [ceph-users] OSD scrub during recovery

2017-05-30 Thread Wido den Hollander
> Op 30 mei 2017 om 17:37 schreef Reed Dier : > > > Lost an OSD and having to rebuild it. > > 8TB drive, so it has to backfill a ton of data. > Been taking a while, so looked at ceph -s and noticed that deep/scrubs were > running even though I’m running newest Jewel

Re: [ceph-users] Scuttlemonkey signing off...

2017-05-23 Thread Wido den Hollander
Hey Patrick, Thanks for all your work in the last 5 years! Sad to see you leave, but again, your effort is very much appreciated! Wido > Op 22 mei 2017 om 16:36 schreef Patrick McGarry : > > > Hey cephers, > > I'm writing to you today to share that my time in the Ceph

Re: [ceph-users] Cephalocon Cancelled

2017-05-12 Thread Wido den Hollander
> Op 12 mei 2017 om 21:45 schreef Patrick McGarry : > > > Hey cephers, > > Sorry to be the bearer of bad news on a Friday, but the decision was > made this week to cancel the Ceph conference that was planned for > later this year in Boston on 23-25 August. > > For more

Re: [ceph-users] CephFS Performance

2017-05-09 Thread Wido den Hollander
> Op 9 mei 2017 om 20:26 schreef Brady Deetz : > > > If I'm reading your cluster diagram correctly, I'm seeing a 1gbps > interconnect, presumably cat6. Due to the additional latency of performing > metadata operations, I could see cephfs performing at those speeds. Are you >

Re: [ceph-users] CentOS 7 and ipv4 is trying to bind ipv6

2017-05-06 Thread Wido den Hollander
> Op 6 mei 2017 om 20:00 schreef Peter Maloney > : > > > I was testing ceph in a centos 7 VM, just so I could support someone > that wants to use CentOS 7, and I ran into a problem... What am I doing > wrong? > > First I tried just without any ipv6 changes,

Re: [ceph-users] jewel - rgw blocked on deep-scrub of bucket index pg

2017-05-06 Thread Wido den Hollander
> Op 6 mei 2017 om 9:55 schreef Christian Balzer <ch...@gol.com>: > > > > Hello, > > On Sat, 6 May 2017 09:25:15 +0200 (CEST) Wido den Hollander wrote: > > > > Op 5 mei 2017 om 10:33 schreef Sam Wouters <s...@ericom.be>: > > > > &g

Re: [ceph-users] RGW: removal of support for fastcgi

2017-05-06 Thread Wido den Hollander
> Op 5 mei 2017 om 21:51 schreef Yehuda Sadeh-Weinraub : > > > RGW has supported since forever. Originally it was the only supported > frontend, and nowadays it is the least preferred one. > > Rgw was first developed over fastcgi + lighttpd, but there were some > issues with

Re: [ceph-users] jewel - rgw blocked on deep-scrub of bucket index pg

2017-05-06 Thread Wido den Hollander
> Op 5 mei 2017 om 10:33 schreef Sam Wouters : > > > Hi, > > we have a small cluster running on jewel 10.2.7; NL-SAS disks only, osd > data and journal co located on the disks; main purpose rgw secondary zone. > > Since the upgrade to jewel, whenever a deep scrub starts on one

Re: [ceph-users] osd and/or filestore tuning for ssds?

2017-05-01 Thread Wido den Hollander
> Op 28 april 2017 om 19:14 schreef Sage Weil : > > > Hi everyone, > > Are there any osd or filestore options that operators are tuning for > all-SSD clusters? If so (and they make sense) we'd like to introduce them > as defaults for ssd-backed OSDs. > osd_op_threads and

Re: [ceph-users] Maintaining write performance under a steady intake of small objects

2017-04-26 Thread Wido den Hollander
> Op 24 april 2017 om 19:52 schreef Florian Haas : > > > Hi everyone, > > so this will be a long email — it's a summary of several off-list > conversations I've had over the last couple of weeks, but the TL;DR > version is this question: > > How can a Ceph cluster

Re: [ceph-users] ceph packages on stretch from eu.ceph.com

2017-04-26 Thread Wido den Hollander
> Op 25 april 2017 om 20:07 schreef Ronny Aasen : > > > Hello > > i am trying to install ceph on debian stretch from > > http://eu.ceph.com/debian-jewel/dists/ > > but there is no stretch repo there. > > now with stretch being frozen, it is a good time to be

[ceph-users] PHP client for RGW Admin Ops API

2017-04-18 Thread Wido den Hollander
Hi, I wanted to share a PHP client for the RGW Admin Ops API [0] which has been developed at my company. There is a proper Python [1] client for the API, but we were unable to find one for PHP, so we wrote it: https://github.com/PCextreme/rgw-admin-php The client works with PHP 7 and is a

Re: [ceph-users] python3-rados

2017-04-10 Thread Wido den Hollander
> Op 8 april 2017 om 4:03 schreef Gerald Spencer : > > > Do the rados bindings exist for python3? > I see this sprinkled in various areas.. > https://github.com/ceph/ceph/pull/7621 > https://github.com/ceph/ceph/blob/master/debian/python3-rados.install > > This being

Re: [ceph-users] best way to resolve 'stale+active+clean' after disk failure

2017-04-06 Thread Wido den Hollander
> Op 7 april 2017 om 1:04 schreef Ben Hines : > > > Personally before extreme measures like marking lost, i would try bringing > up the osd, so it's up and out -- i believe the data will still be found > and re balanced away from it by Ceph. Indeed, do not mark it as lost

Re: [ceph-users] Apply for an official mirror at CN

2017-04-05 Thread Wido den Hollander
> On Sat, Apr 1, 2017 at 8:10 PM, Wido den Hollander <w...@42on.com> wrote: > >> Great! Very good to hear. We can CNAME cn.ceph.com to that location? > > > > > > Yes, please CNAME to mirrors.ustc.edu.cn, and I will set vhost in our > > nginx for the &

Re: [ceph-users] FreeBSD port net/ceph-devel released

2017-04-01 Thread Wido den Hollander
> Op 31 maart 2017 om 19:15 schreef Willem Jan Withagen <w...@digiware.nl>: > > > On 31-3-2017 17:32, Wido den Hollander wrote: > > Hi Willem Jan, > > > >> Op 30 maart 2017 om 13:56 schreef Willem Jan Withagen > >> <w...@digiware.nl>: &

Re: [ceph-users] clock skew

2017-04-01 Thread Wido den Hollander
ohibited. If you received > this in error, please contact the sender and delete the material from any > computer. > > On Sat, Apr 1, 2017 at 9:12 AM, mj <li...@merit.unu.edu> wrote: > > > Hi, > > > > On 04/01/2017 02:10 PM, Wido den Hollander wrote: > >

Re: [ceph-users] Apply for an official mirror at CN

2017-04-01 Thread Wido den Hollander
> Op 1 april 2017 om 12:20 schreef SJ Zhu : > > > Hi all, > > We'd like to apply for the official mirror at China. > We have run the ceph mirror for more than one year, at > https://mirrors.ustc.edu.cn/ceph/ > > The server infomation is: > * located at USTC(University of

Re: [ceph-users] clock skew

2017-04-01 Thread Wido den Hollander
> Op 1 april 2017 om 11:17 schreef mj : > > > Hi, > > Despite ntp, we keep getting clock skews that auto disappear again after > a few minutes. > > To prevent the unneccerasy HEALTH_WARNs, I have increased the > mon_clock_drift_allowed to 0.2, as can be seen below: >

Re: [ceph-users] FreeBSD port net/ceph-devel released

2017-03-31 Thread Wido den Hollander
Hi Willem Jan, > Op 30 maart 2017 om 13:56 schreef Willem Jan Withagen : > > > Hi, > > I'm pleased to announce that my efforts to port to FreeBSD have resulted > in a ceph-devel port commit in the ports tree. > > https://www.freshports.org/net/ceph-devel/ > Awesome work! I

Re: [ceph-users] cephfs and erasure coding

2017-03-29 Thread Wido den Hollander
> Op 29 maart 2017 om 8:54 schreef Konstantin Shalygin : > > > Hello. > > How your tests? I'm looking for CephFS with EC for save space on > replicas for many small files (dovecot mailboxes). I wouldn't use CephFS for so many small files. Dovecot will do a lot of locking,

Re: [ceph-users] At what point are objects removed?

2017-03-28 Thread Wido den Hollander
> Op 28 maart 2017 om 16:52 schreef Gregory Farnum : > > > CephFS files are deleted asynchronously by the mds once there are no more > client references to the file (NOT when the file is unlinked -- that's not > how posix works). If the number of objects isn't going down

Re: [ceph-users] Ceph OSD network with IPv6 SLAAC networks?

2017-03-28 Thread Wido den Hollander
> Op 27 maart 2017 om 21:49 schreef Richard Hesse : > > > Has anyone run their Ceph OSD cluster network on IPv6 using SLAAC? I know > that ceph supports IPv6, but I'm not sure how it would deal with the > address rotation in SLAAC, permanent vs outgoing address, etc.

Re: [ceph-users] New hardware for OSDs

2017-03-27 Thread Wido den Hollander
> Op 27 maart 2017 om 13:22 schreef Christian Balzer : > > > > Hello, > > On Mon, 27 Mar 2017 12:27:40 +0200 Mattia Belluco wrote: > > > Hello all, > > we are currently in the process of buying new hardware to expand an > > existing Ceph cluster that already has 1200 osds. >

Re: [ceph-users] leveldb takes a lot of space

2017-03-27 Thread Wido den Hollander
> Op 26 maart 2017 om 9:44 schreef Niv Azriel : > > > after network issues, ceph cluster fails. > leveldb grows and takes a lot of space > ceph mon cant write to leveldb because there is not enough space on > filesystem. > (there is a lot of ldb file on /var/lib/ceph/mon) >

Re: [ceph-users] New metrics.ceph.com!

2017-03-23 Thread Wido den Hollander
> Op 22 maart 2017 om 18:05 schreef Patrick McGarry : > > > Hey cephers, > > Just wanted to share that the new interactive metrics dashboard is now > available for tire-kicking. > > https://metrics.ceph.com > Very nice! > There are still a few data pipeline issues and

Re: [ceph-users] OSDs cannot match up with fast OSD map changes (epochs) during recovery

2017-03-20 Thread Wido den Hollander
628206 7f2056735700 0 -- > 10.139.4.69:6836/777372 >> - conn(0x7f20c1bfa800 :6836 > s=STATE_ACCEPTING_WAIT_BANNER_ADDR pgs=0 cs=0 l=0).fault with nothing to > send and in the half accept state just closed messages on many osds which > are recovering. > > Suggestions woul

Re: [ceph-users] Ceph-osd Daemon Receives Segmentation Fault on Trusty After Upgrading to 0.94.10 Release

2017-03-19 Thread Wido den Hollander
> Op 17 maart 2017 om 8:39 schreef Özhan Rüzgar Karaman > : > > > Hi; > Yesterday i started to upgrade my Ceph environment from 0.94.9 to 0.94.10. > All monitor servers upgraded successfully but i experience problems on > starting upgraded OSD daemons. > > When i try

Re: [ceph-users] modify civetweb default port won't work

2017-03-13 Thread Wido den Hollander
> Op 13 maart 2017 om 15:03 schreef Yair Magnezi : > > > Hello Cephers . > > I'm trying to modify the civetweb default port to 80 but from some > reason it insists on listening on the default 7480 port > > My configuration is quiet simple ( experimental ) and

Re: [ceph-users] MySQL and ceph volumes

2017-03-08 Thread Wido den Hollander
> Op 8 maart 2017 om 0:35 schreef Matteo Dacrema : > > > Thank you Adrian! > > I’ve forgot this option and I can reproduce the problem. > > Now, what could be the problem on ceph side with O_DSYNC writes? > As mentioned nothing, but what you can do with MySQL is provide

<    1   2   3   4   5   6   7   8   9   10   >