Re: [ceph-users] Best practices for allocating memory to bluestore cache

2018-08-30 Thread Bastiaan Visser
Your claim that all cache is used for K/V cache is false (with default settings). K/V cache is maximized to 500Meg. : bluestore_cache_kv_max Description:The maximum amount of cache devoted to key/value data (rocksdb). Type: Unsigned Integer Required: Yes Default:512 *

[ceph-users] MDS not start. Timeout??

2018-08-30 Thread morf...@gmail.com
Hello all! I had a electric power problem. After this I have 2 incomplete pg. But all RBD volumes are work. But not work my CephFS. MDS load stop at "replay" state and MDS related commands hangs: cephfs-journal-tool journal export backup.bin - freeze; cephfs-journal-tool event

Re: [ceph-users] Strange Client admin socket error in a containerized ceph environment

2018-08-30 Thread Alex Litvak
Also, I tried to add a client admin socket, but I am getting this message: [client] admin socket = /var/run/ceph/$name.$pid.asok 2018-08-30 20:04:03.337 7f4641ffb700 -1 set_mon_vals failed to set admin_socket = $run_dir/$cluster-$name.asok: Configuration option 'admin_socket' may not be

[ceph-users] Strange Client admin socket error in a containerized ceph environment

2018-08-30 Thread Alex Litvak
I keep getting the following error message: 018-08-30 18:52:37.882 7fca9df7c700 -1 asok(0x7fca98000fe0) AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to '/var/run/ceph/ceph-client.admin.asok': (17) File exists Otherwise things seem

Re: [ceph-users] Best practices for allocating memory to bluestore cache

2018-08-30 Thread David Turner
Be very careful trying to utilize more RAM while your cluster is healthy. When you're going to need the extra RAM is when you're closer is unhealthy and your osds are peering, recovering, backfilling, etc. That's when the osd daemons start needing the RAM that is recommended in the docs. On Thu,

[ceph-users] Best practices for allocating memory to bluestore cache

2018-08-30 Thread Tyler Bishop
Hi, My OSD host has 256GB of ram and I have 52 OSD. Currently I have the cache set to 1GB and the system only consumes around 44GB of ram and the other ram sits as unallocated because I am using bluestore vs filestore. The documentation:

Re: [ceph-users] Luminous missing osd_backfill_full_ratio

2018-08-30 Thread David Turner
This moved to the PG map in luminous. I think it might have been there in Jewel as well. http://docs.ceph.com/docs/luminous/man/8/ceph/#pg ceph pg set_full_ratio ceph pg set_backfillfull_ratio ceph pg set_nearfull_ratio On Thu, Aug 30, 2018, 1:57 PM David C wrote: > Hi All > > I feel like

[ceph-users] Luminous missing osd_backfill_full_ratio

2018-08-30 Thread David C
Hi All I feel like this is going to be a silly query with a hopefully simple answer. I don't seem to have the osd_backfill_full_ratio config option on my OSDs and can't inject it. This a Lumimous 12.2.1 cluster that was upgraded from Jewel. I added an OSD to the cluster and woke up the next day

Re: [ceph-users] MDS does not always failover to hot standby on reboot

2018-08-30 Thread Gregory Farnum
LOn Thu, Aug 30, 2018 at 12:46 PM William Lawton wrote: Oh i see. We’d taken steps to reduce the risk of losing the active mds and > mon leader instances at the same time in the hope that it would prevent > this issue. Do you know if the mds always connects to a specific mon > instance i.e. the

Re: [ceph-users] MDS does not always failover to hot standby on reboot

2018-08-30 Thread William Lawton
Oh i see. We’d taken steps to reduce the risk of losing the active mds and mon leader instances at the same time in the hope that it would prevent this issue. Do you know if the mds always connects to a specific mon instance i.e. the mon provider and can it be determined which mon instance that

Re: [ceph-users] Ceph-Deploy error on 15/71 stage

2018-08-30 Thread Jones de Andrade
Hi Eugen. Ok, edited the file /etc/salt/minion, uncommented the "log_level_logfile" line and set it to "debug" level. Turned off the computer, waited a few minutes so that the time frame would stand out in the /var/log/messages file, and restarted the computer. Using vi I "greped out" (awful

Re: [ceph-users] MDS does not always failover to hot standby on reboot

2018-08-30 Thread Gregory Farnum
Okay, well that will be the same reason then. If the active MDS is connected to a monitor and they fail at the same time, the monitors can’t replace the mds until they’ve been through their own election and a full mds timeout window. On Thu, Aug 30, 2018 at 11:46 AM William Lawton wrote: >

[ceph-users] SPDK/DPDK with Intel P3700 NVMe pool

2018-08-30 Thread Kevin Olbrich
Hi! During our move from filestore to bluestore, we removed several Intel P3700 NVMe from the nodes. Is someone running a SPDK/DPDK NVMe-only EC pool? Is it working well? The docs are very short about the setup:

Re: [ceph-users] rocksdb mon stores growing until restart

2018-08-30 Thread David Turner
The Hammer ticket was https://tracker.ceph.com/issues/13990. The problem here was when OSDs asked each other for which map they needed to keep and a leak would set it to NULL then that OSD would never delete an OSD map again until it was restarted. On Thu, Aug 30, 2018 at 3:09 AM Joao Eduardo

Re: [ceph-users] MDS does not always failover to hot standby on reboot

2018-08-30 Thread William Lawton
Thanks for the response Greg. We did originally have co-located mds and mon but realised this wasn't a good idea early on and separated them out onto different hosts. So our mds hosts are on ceph-01 and ceph-02, and our mon hosts are on ceph-03, 04 and 05. Unfortunately we see this issue

Re: [ceph-users] mimic/bluestore cluster can't allocate space for bluefs

2018-08-30 Thread David Turner
I have 2 OSDs failing to start due to this [1] segfault. What is happening matches what Sage said about this [2] bug. The OSDs are NVMe disks and rocksdb is compacting omaps. I attempted setting `bluestore_bluefs_min_free = 10737418240` and then start the OSDs, but they both segfaulted with the

Re: [ceph-users] cephfs speed

2018-08-30 Thread Peter Eisch
Thanks for the thought. It’s mounted with this entry in fstab (one line, if email wraps it): cephmon-s01,cephmon-s02,cephmon-s03:/     /loam    ceph    noauto,name=clientname,secretfile=/etc/ceph/secret,noatime,_netdev    0       2 Pretty plain, but I'm open to tweaking! peter Peter Eisch

Re: [ceph-users] safe to remove leftover bucket index objects

2018-08-30 Thread David Turner
I'm glad you asked this, because it was on my to-do list. I know that based on our not existing in the bucket marker does not mean it's safe to delete. I have an index pool with 22k objects in it. 70 objects match existing bucket markers. I was having a problem on the cluster and started deleting

Re: [ceph-users] SSD OSDs crashing after upgrade to 12.2.7

2018-08-30 Thread Alfredo Deza
On Thu, Aug 30, 2018 at 5:24 AM, Wolfgang Lendl wrote: > Hi Alfredo, > > > caught some logs: > https://pastebin.com/b3URiA7p That looks like there is an issue with bluestore. Maybe Radoslaw or Adam might know a bit more. > > br > wolfgang > > On 2018-08-29 15:51, Alfredo Deza wrote: >> On Wed,

Re: [ceph-users] cephfs speed

2018-08-30 Thread Gregory Farnum
How are you mounting CephFS? It may be that the cache settings are just set very badly for a 10G pipe. Plus rados bench is a very parallel large-IO benchmark and many benchmarks you might dump into a filesystem are definitely not. -Greg On Thu, Aug 30, 2018 at 7:54 AM Peter Eisch wrote: > Hi, >

Re: [ceph-users] MDS does not always failover to hot standby on reboot

2018-08-30 Thread Gregory Farnum
Yes, this is a consequence of co-locating the MDS and monitors — if the MDS reports to its co-located monitor and both fail, the monitor cluster has to go through its own failure detection and then wait for a full MDS timeout period after that before it marks the MDS down. :( We might conceivably

[ceph-users] MDS does not always failover to hot standby on reboot

2018-08-30 Thread William Lawton
Hi. We have a 5 node Ceph cluster (refer to ceph -s output at bottom of email). During resiliency tests we have an occasional problem when we reboot the active MDS instance and a MON instance together i.e. dub-sitv-ceph-02 and dub-sitv-ceph-04. We expect the MDS to failover to the standby

[ceph-users] cephfs speed

2018-08-30 Thread Peter Eisch
Hi, I have a cluster serving cephfs and it works. It’s just slow. Client is using the kernel driver. I can ‘rados bench’ writes to the cephfs_data pool at wire speeds (9580Mb/s on a 10G link) but when I copy data into cephfs it is rare to get above 100Mb/s. Large file writes may start fast

Re: [ceph-users] Clients report OSDs down/up (dmesg) nothing in Ceph logs (flapping OSDs)

2018-08-30 Thread Eugen Block
Correct, except it doesn't have to be a specific host or a specific OSD. What matters here is whether the client is idle. As soon as the client is woken up and sends a request to _any_ OSD, it receives a new osdmap and applies it, possibly emitting those dmesg entries. Thanks for the

Re: [ceph-users] Clients report OSDs down/up (dmesg) nothing in Ceph logs (flapping OSDs)

2018-08-30 Thread Ilya Dryomov
On Thu, Aug 30, 2018 at 1:04 PM Eugen Block wrote: > > Hi again, > > we still didn't figure out the reason for the flapping, but I wanted > to get back on the dmesg entries. > They just reflect what happened in the past, they're no indicator to > predict anything. The kernel client is just that,

Re: [ceph-users] safe to remove leftover bucket index objects

2018-08-30 Thread Dan van der Ster
Replying to self... On Wed, Aug 1, 2018 at 11:56 AM Dan van der Ster wrote: > > Dear rgw friends, > > Somehow we have more than 20 million objects in our > default.rgw.buckets.index pool. > They are probably leftover from this issue we had last year: >

Re: [ceph-users] librmb: Mail storage on RADOS with Dovecot

2018-08-30 Thread Marc Roos
How is it going with this? Are we getting close to a state where we can store a mailbox on ceph with this librmb? -Original Message- From: Wido den Hollander [mailto:w...@42on.com] Sent: maandag 25 september 2017 9:20 To: Gregory Farnum; Danny Al-Gaaf Cc: ceph-users Subject: Re:

Re: [ceph-users] SSD OSDs crashing after upgrade to 12.2.7

2018-08-30 Thread Wolfgang Lendl
Hi Alfredo, caught some logs: https://pastebin.com/b3URiA7p br wolfgang On 2018-08-29 15:51, Alfredo Deza wrote: > On Wed, Aug 29, 2018 at 2:06 AM, Wolfgang Lendl > wrote: >> Hi, >> >> after upgrading my ceph clusters from 12.2.5 to 12.2.7 I'm experiencing >> random crashes from SSD OSDs

Re: [ceph-users] rocksdb mon stores growing until restart

2018-08-30 Thread Joao Eduardo Luis
On 08/30/2018 09:28 AM, Dan van der Ster wrote: > Hi, > > Is anyone else seeing rocksdb mon stores slowly growing to >15GB, > eventually triggering the 'mon is using a lot of disk space' warning? > > Since upgrading to luminous, we've seen this happen at least twice. > Each time, we restart all

[ceph-users] CephFS : fuse client vs kernel driver

2018-08-30 Thread Hervé Ballans
Hi all, I just finished setting up a new Ceph cluster (Luminous 12.2.7, 3xMON nodes and 6xOSDs nodes, BlueStore OSD on sata hdd with WAL/DB on separated NVMe devices, 2x10 Gbs network per node, 3 replicas by pool) I created a CephFS pool : data pool uses hdd OSDs and metadata pool uses

Re: [ceph-users] Clients report OSDs down/up (dmesg) nothing in Ceph logs (flapping OSDs)

2018-08-30 Thread Eugen Block
Hi again, we still didn't figure out the reason for the flapping, but I wanted to get back on the dmesg entries. They just reflect what happened in the past, they're no indicator to predict anything. For example, when I changed the primary-affinity of OSD.24 last week, one of the clients

[ceph-users] rocksdb mon stores growing until restart

2018-08-30 Thread Dan van der Ster
Hi, Is anyone else seeing rocksdb mon stores slowly growing to >15GB, eventually triggering the 'mon is using a lot of disk space' warning? Since upgrading to luminous, we've seen this happen at least twice. Each time, we restart all the mons and then stores slowly trim down to <500MB. We have

[ceph-users] Odp.: New Ceph community manager: Mike Perez

2018-08-30 Thread Tomasz Kuzemko
Welcome Mike! You're the perfect person for this role! -- Tomasz Kuzemko tomasz.kuze...@corp.ovh.com Od: ceph-users w imieniu użytkownika Sage Weil Wysłane: środa, 29 sierpnia 2018 03:13 Do: ceph-de...@vger.kernel.org; ceph-users@lists.ceph.com;

Re: [ceph-users] Ceph-Deploy error on 15/71 stage

2018-08-30 Thread Eugen Block
Hi, So, it only contains logs concerning the node itself (is it correct? sincer node01 is also the master, I was expecting it to have logs from the other too) and, moreover, no ceph-osd* files. Also, I'm looking the logs I have available, and nothing "shines out" (sorry for my poor english) as