Re: [ceph-users] ceph 12.2.5 - atop DB/WAL SSD usage 0%

2018-04-30 Thread Hans van den Bogert
Shouldn't Steven see some data being written to the block/wal for object metadata? Though that might be negligible with 4MB objects On 27-04-18 16:04, Serkan Çoban wrote: rados bench is using 4MB block size for io. Try with with io size 4KB, you will see ssd will be used for write

Re: [ceph-users] ceph luminous 12.2.4 - 2 servers better than 3 ?

2018-04-19 Thread Hans van den Bogert
id: 1): Disk Write Cache : Disk's Default > Adapter 0-VD 2(target id: 2): Disk Write Cache : Disk's Default > Adapter 0-VD 3(target id: 3): Disk Write Cache : Disk's Default > > > On Thu, 19 Apr 2018 at 14:22, Hans van den Bogert <hansbog...@gmail.com> > wrote: > >>

Re: [ceph-users] ceph luminous 12.2.4 - 2 servers better than 3 ?

2018-04-19 Thread Hans van den Bogert
cation of WAL/DB ( on separate SSD or same HDD) Thanks Steven On Thu, 19 Apr 2018 at 12:06, Hans van den Bogert <hansbog...@gmail.com> wrote: > I take it that the first bench is with replication size 2, the second > bench is with replication size 3? Same for the 4 node OSD scenario?

Re: [ceph-users] ceph luminous 12.2.4 - 2 servers better than 3 ?

2018-04-19 Thread Hans van den Bogert
04 > Object size:4194304 > Bandwidth (MB/sec): 44.0793 > Stddev Bandwidth: 55.3843 > Max bandwidth (MB/sec): 232 > Min bandwidth (MB/sec): 0 > Average IOPS: 11 > Stddev IOPS:13 > Max IOPS: 58 > Min IOPS: 0 &

Re: [ceph-users] ceph luminous 12.2.4 - 2 servers better than 3 ?

2018-04-19 Thread Hans van den Bogert
Hi Steven, There is only one bench. Could you show multiple benches of the different scenarios you discussed? Also provide hardware details. Hans On Apr 19, 2018 13:11, "Steven Vacaroaia" wrote: Hi, Any idea why 2 servers with one OSD each will provide better performance

Re: [ceph-users] scalability new node to the existing cluster

2018-04-18 Thread Hans van den Bogert
I keep seeing these threads where adding nodes has such an impact on the cluster as a whole, that I wonder what the rest of the cluster looks like. Normally I’d just advise someone to put a limit on the concurrent backfills that can be done, and `osd max backfills` by default already is 1.

Re: [ceph-users] Luminous 12.2.3 release date?

2018-02-12 Thread Hans van den Bogert
Hi Wido, Did you ever get an answer? I'm eager to know as well. Hans On Tue, Jan 30, 2018 at 10:35 AM, Wido den Hollander wrote: > Hi, > > Is there a ETA yet for 12.2.3? Looking at the tracker there aren't that many > outstanding issues:

[ceph-users] Retrieving ceph health from restful manager plugin

2018-02-05 Thread Hans van den Bogert
Hi All, I might really be bad at searching, but I can't seem to find the ceph health status through the new(ish) restful api. Is that right? I know how I could retrieve it through a Python script, however I'm trying to keep our monitoring application as layer cake free as possible -- as such a

[ceph-users] Redirect for restful API in manager

2018-02-05 Thread Hans van den Bogert
Hi all, In the release notes of 12.2.2 the following is stated: > Standby ceph-mgr daemons now redirect requests to the active messenger, easing configuration for tools & users accessing the web dashboard, restful API, or other ceph-mgr module services. However, it doesn't seem to be the

Re: [ceph-users] Fwd: Ceph team involvement in Rook (Deploying Ceph in Kubernetes)

2018-01-21 Thread Hans van den Bogert
Should I summarize this is ceph-helm being being EOL? If I'm spinning up a toy cluster for a homelab, should I invest time in Rook, or stay with ceph-helm for now? On Fri, Jan 19, 2018 at 11:55 AM, Kai Wagner wrote: > Just for those of you who are not subscribed to ceph-users.

Re: [ceph-users] Increasing PG number

2018-01-02 Thread Hans van den Bogert
Please refer to standard documentation as much as possible, http://docs.ceph.com/docs/jewel/rados/operations/placement-groups/#set-the-number-of-placement-groups Han’s is also

Re: [ceph-users] The way to minimize osd memory usage?

2017-12-11 Thread Hans van den Bogert
There’s probably multiple reasons. However I just wanted to chime in that I set my cache size to 1G and I constantly see OSD memory converge to ~2.5GB. In [1] you can see the difference between a node with 4 OSDs, v12.2.2, on the left; and a node with 4 OSDs v12.2.1 on the right. I really

[ceph-users] osd/bluestore: Get block.db usage

2017-12-04 Thread Hans van den Bogert
Hi all, Is there a way to get the current usage of the bluestore's block.db? I'd really like to monitor this as we have a relatively high number of objects per OSD. A second question related to the above, are there mechanisms to influence which objects' metadata gets spilled once the block.db is

Re: [ceph-users] ceps-deploy won't install luminous

2017-11-15 Thread Hans van den Bogert
verify that you did that part? > On Nov 15, 2017, at 10:41 AM, Hans van den Bogert <hansbog...@gmail.com> > wrote: > > Hi, > > Can you show the contents of the file, /etc/yum.repos.d/ceph.repo ? > > Regards, > > Hans >> On Nov 15, 2017, at 10:27 AM,

Re: [ceph-users] ceps-deploy won't install luminous

2017-11-15 Thread Hans van den Bogert
Hi, Can you show the contents of the file, /etc/yum.repos.d/ceph.repo ? Regards, Hans > On Nov 15, 2017, at 10:27 AM, Ragan, Tj (Dr.) > wrote: > > Hi All, > > I feel like I’m doing something silly. I’m spinning up a new cluster, and > followed the instructions on

Re: [ceph-users] Fwd: Luminous RadosGW issue

2017-11-09 Thread Hans van den Bogert
AC24O8CJ0A9Gn6qeiPalEYg== > caps mon = "allow rwx" > caps osd = "allow rwx" > > > 2017-11-09 6:11 GMT+07:00 Hans van den Bogert <hansbog...@gmail.com > <mailto:hansbog...@gmail.com>>: > Are you sure you deployed it with

Re: [ceph-users] Fwd: Luminous RadosGW issue

2017-11-08 Thread Hans van den Bogert
Are you sure you deployed it with the client.radosgw.gateway name as well? Try to redeploy the RGW and make sure the name you give it corresponds to the name you give in the ceph.conf. Also, do not forget to push the ceph.conf to the RGW machine. On Wed, Nov 8, 2017 at 11:44 PM, Sam Huracan

Re: [ceph-users] Ceph versions not showing RGW

2017-11-02 Thread Hans van den Bogert
Just to get this really straight, Jewel OSDs do send this metadata? Otherwise I'm probably mistaken that I ever saw 10.2.x versions in the output. Thanks, Hans On 2 Nov 2017 12:31 PM, "John Spray" <jsp...@redhat.com> wrote: > On Thu, Nov 2, 2017 at 11:16 AM, Hans van de

[ceph-users] Ceph versions not showing RGW

2017-11-02 Thread Hans van den Bogert
Hi all, During our upgrade from Jewel to Luminous I saw the following behaviour, if my memory serves me right: When upgrading for example monitors and OSDs, we saw that the `ceph versions` command correctly showed at one that some OSDs were still on Jewel (10.2.x) and some were already upgraded

Re: [ceph-users] PGs inconsistent, do I fear data loss?

2017-11-02 Thread Hans van den Bogert
Never mind, I should’ve read the whole thread first. > On Nov 2, 2017, at 10:50 AM, Hans van den Bogert <hansbog...@gmail.com> wrote: > > >> On Nov 1, 2017, at 4:45 PM, David Turner <drakonst...@gmail.com >> <mailto:drakonst...@gmail.com>> wrote: >>

Re: [ceph-users] PGs inconsistent, do I fear data loss?

2017-11-02 Thread Hans van den Bogert
> On Nov 1, 2017, at 4:45 PM, David Turner wrote: > > All it takes for data loss is that an osd on server 1 is marked down and a > write happens to an osd on server 2. Now the osd on server 2 goes down > before the osd on server 1 has finished backfilling and the first

Re: [ceph-users] announcing ceph-helm (ceph on kubernetes orchestration)

2017-10-25 Thread Hans van den Bogert
Very interesting. I've been toying around with Rook.io [1]. Did you know of this project, and if so can you tell if ceph-helm and Rook.io have similar goals? Regards, Hans [1] https://rook.io/ On 25 Oct 2017 21:09, "Sage Weil" wrote: > There is a new repo under the ceph

[ceph-users] Drive write cache recommendations for Luminous/Bluestore

2017-10-23 Thread Hans van den Bogert
Hi All, For Jewel there is this page about drive cache: http://docs.ceph.com/docs/jewel/rados/configuration/filesystem-recommendations/#hard-drive-prep For Bluestore I can't find any documentation or discussions about drive write cache, while I can imagine that revisiting this subject might be

Re: [ceph-users] Ceph delete files and status

2017-10-20 Thread Hans van den Bogert
My experience with RGW is that actual freeing up of space is asynchronous to the a S3 client’s command to delete an object. I.e., it might take a while before it’s actually freed up. Can you redo your little experiment and simply wait for an hour to let the garbage collector to do its thing, or

Re: [ceph-users] High mem with Luminous/Bluestore

2017-10-19 Thread Hans van den Bogert
> Memory usage is still quite high here even with a large onode cache! > Are you using erasure coding? I recently was able to reproduce a bug in > bluestore causing excessive memory usage during large writes with EC, > but have not tracked down exactly what's going on yet. > > Mark No, this

Re: [ceph-users] High mem with Luminous/Bluestore

2017-10-18 Thread Hans van den Bogert
d monitor the memory usage. Thanks, Hans On Wed, Oct 18, 2017 at 11:56 AM, Wido den Hollander <w...@42on.com> wrote: > > > Op 18 oktober 2017 om 11:41 schreef Hans van den Bogert < > hansbog...@gmail.com>: > > > > > > Hi All, > > > > I'

[ceph-users] High mem with Luminous/Bluestore

2017-10-18 Thread Hans van den Bogert
Hi All, I've converted 2 nodes with 4 HDD/OSDs each from Filestore to Bluestore. I expected somewhat higher memory usage/RSS values, however I see, imo, a huge memory usage for all OSDs on both nodes. Small snippet from `top` PID USER PR NIVIRTRESSHR S %CPU %MEM TIME+

[ceph-users] How to get current min-compat-client setting

2017-10-13 Thread Hans van den Bogert
Hi, I’m in the middle of debugging some incompatibilities with an upgrade of Proxmox which uses Ceph. At this point I’d like to know what my current value is for the min-compat-client setting, which would’ve been set by: ceph osd set-require-min-compat-client … AFAIK, there is no direct

Re: [ceph-users] Gracefully reboot OSD node

2017-08-03 Thread Hans van den Bogert
, 2017 at 1:55 PM, Hans van den Bogert <hansbog...@gmail.com> wrote: > What are the implications of this? Because I can see a lot of blocked > requests piling up when using 'noout' and 'nodown'. That probably makes > sense though. > Another thing, no when the OSDs come back o

Re: [ceph-users] Gracefully reboot OSD node

2017-08-03 Thread Hans van den Bogert
, 2017 at 1:36 PM, linghucongsong <linghucongs...@163.com> wrote: > > > set the osd noout nodown > > > > > At 2017-08-03 18:29:47, "Hans van den Bogert" <hansbog...@gmail.com> > wrote: > > Hi all, > > One thing which has bothered since t

[ceph-users] Gracefully reboot OSD node

2017-08-03 Thread Hans van den Bogert
Hi all, One thing which has bothered since the beginning of using ceph is that a reboot of a single OSD causes a HEALTH_ERR state for the cluster for at least a couple of seconds. In the case of planned reboot of a OSD node, should I do some extra commands in order not to go to HEALTH_ERR state?

[ceph-users] Linear space complexity or memory leak in `Radosgw-admin bucket check --fix`

2017-07-25 Thread Hans van den Bogert
Hi All, I don't seem to be able to fix a bucket, a bucket which has become inconsistent due to the use of the `inconsistent-index` flag 8). My ceph-admin VM has 4GB of RAM, but that doesn't seem to be enough to do a `radosgw-admin bucket check --fix` which holds 6M items, as the radosgw-admin

[ceph-users] Crash on startup

2017-02-01 Thread Hans van den Bogert
Hi All, I'm clueless as to why an OSD crashed. I have a log at [1]. If anyone can explain how this should be interpreted, then please let me know. I can only see generic errors probably started by a false assert. Restarting the OSD fails with the same errors as in [1]. It seems like, though