[ceph-users] Several OSDs won't come up. Worried for complete data loss

2020-01-22 Thread Justin Engwer
Hi, I've recently upgraded from Nautilus 14.2.2 to 14.2.6. I've also been installing some new OSDs to my cluster. It looks as though either the backplane I've added has power issues or the raid card I've added has bad memory. Several new-ish, known good drives were bounced out of their JBOD

Re: [ceph-users] Problem : "1 pools have many more objects per pg than average"

2020-01-22 Thread Nathan Fish
Injectargs causes an immediate runtime change; rebooting the mon would negate the change. On Wed., Jan. 22, 2020, 4:41 p.m. St-Germain, Sylvain (SSC/SPC), < sylvain.st-germ...@canada.ca> wrote: > / Problem /// > > I've got a Warning on my cluster

[ceph-users] Problem : "1 pools have many more objects per pg than average"

2020-01-22 Thread St-Germain, Sylvain (SSC/SPC)
/ Problem /// I've got a Warning on my cluster that I cannot remove : "1 pools have many more objects per pg than average" Does somebody has some insight ? I think it's normal to have this warning because I have just one pool in use, but how can

Re: [ceph-users] S3 Bucket usage up 150% diference between rgw-admin and external metering tools.

2020-01-22 Thread Robin H. Johnson
On Tue, Jan 21, 2020 at 05:57:51PM +, Robin H. Johnson wrote: > On Mon, Jan 20, 2020 at 12:57:51PM +, EDH - Manuel Rios wrote: > > Hi Cephs > > > > Several nodes of our Ceph 14.2.5 are fully dedicated to host cold storage / > > backups information. > > > > Today checking the data usage

[ceph-users] Rados bench behaves oddly

2020-01-22 Thread John Hearns
We have a CEPH storage cluster which is having problems. When I run a rados bench I get the behaviour below. Has anyone seen this sort of thing before? # rados bench -p scbench 10 seq hints = 1 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 0 0 0

Re: [ceph-users] CephFS with cache-tier kernel-mount client unable to write (Nautilus)

2020-01-22 Thread Hayashida, Mami
Thanks, Ilya. I just tried modifying the osd cap for client.testuser by getting rid of "tag cephfs data=cephfs_test" part and confirmed this key does work (i.e. lets the CephFS client read/write). It now reads: [client.testuser] key = XXXZZZ caps mds = "allow rw" caps mon = "allow r" caps

Re: [ceph-users] OSD crash after change of osd_memory_target

2020-01-22 Thread Igor Fedotov
Hi Martin, looks like a bug to me. You might want to remove all custom settings from config database and try to set osd-memory-target only. Would it help? Thanks, Igor On 1/22/2020 3:43 PM, Martin Mlynář wrote: Dne 21. 01. 20 v 21:12 Stefan Kooman napsal(a): Quoting Martin Mlynář

[ceph-users] Problems with ragosgw

2020-01-22 Thread mohamed zayan
Hi, Currently I have a cluster of 2 nodes on two raspberrypi 3 devices. pi1 is admin/mon/mgr/osd pi2 is osd I am currently trying to run radosgw on pi2. I have failed multiple times /var/lib/ceph/radosgw# /usr/bin/radosgw -f --cluster ceph --name client.rgw.pi2 --setuser ceph --setgroup ceph

Re: [ceph-users] MDS: obscene buffer_anon memory use when scanning lots of files

2020-01-22 Thread John Madden
> Couldn't John confirm that this is the issue by checking the heap stats and > triggering the release via > > ceph tell mds.mds1 heap stats > ceph tell mds.mds1 heap release > > (this would be much less disruptive than restarting the MDS) That was my first thought as well, but `release`

Re: [ceph-users] OSD crash after change of osd_memory_target

2020-01-22 Thread Martin Mlynář
Dne 21. 01. 20 v 21:12 Stefan Kooman napsal(a): > Quoting Martin Mlynář (nexus+c...@smoula.net): > >> Do you think this could help? OSD does not even start, I'm getting a little >> lost how flushing caches could help. > I might have mis-understood. I though the OSDs crashed when you set the >

Re: [ceph-users] Ceph MDS randomly hangs with no useful error message

2020-01-22 Thread Janek Bevendorff
I don't find any clue from the backtrace. please run 'ceph daemon mds. dump_historic_ops' and ''ceph daemon mds.xxx perf reset; ceph daemon mds.xxx perf dump'. send the outputs to us. Hi, I assume you mean ceph daemon mds.xxx perf reset _all_? Here's the output of historic ops

Re: [ceph-users] MDS: obscene buffer_anon memory use when scanning lots of files

2020-01-21 Thread Dan van der Ster
On Wed, Jan 22, 2020 at 12:24 AM Patrick Donnelly wrote: > On Tue, Jan 21, 2020 at 8:32 AM John Madden wrote: > > > > On 14.2.5 but also present in Luminous, buffer_anon memory use spirals > > out of control when scanning many thousands of files. The use case is > > more or less "look up this

[ceph-users] Unable to track different ceph client version connections

2020-01-21 Thread Pardhiv Karri
Hi, We upgraded our Ceph cluster from Hammer to Luminous and it is running fine. Post upgrade we live migrated all our Openstack instances (not 100% sure). Currently we see 1658 clients still on Hammer version. To track the clients we increased the debugging of debug_mon=10/10, debug_ms=1/5,

Re: [ceph-users] S3 Bucket usage up 150% diference between rgw-admin and external metering tools.

2020-01-21 Thread EDH - Manuel Rios
Hi Cbodley , As you requested by IRC we tested directly with AWS Cli. Results: aws --endpoint=http://XX --profile=ceph s3api list-multipart-uploads --bucket Evol6 It reports near 170 uploads. We used the last one: { "Initiator": { "DisplayName": "x",

Re: [ceph-users] MDS: obscene buffer_anon memory use when scanning lots of files

2020-01-21 Thread Patrick Donnelly
On Tue, Jan 21, 2020 at 8:32 AM John Madden wrote: > > On 14.2.5 but also present in Luminous, buffer_anon memory use spirals > out of control when scanning many thousands of files. The use case is > more or less "look up this file and if it exists append this chunk to > it, otherwise create it

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-21 Thread Eric K. Miller
Hi Vitaliy, The drive is a Micron 5200 ECO 3.84TB This is from the msecli utility: Device Name : /dev/sde Model No : Micron_5200_MTFDDAK3T8TDC Serial No: FW-Rev : D1MU404 Total Size : 3840.00GB Drive Status : Drive is in good

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-21 Thread Виталий Филиппов
Hi! Thanks. The parameter gets reset when you reconnect the SSD so in fact it requires not to power cycle it after changing the parameter :-) Ok, this case seems lucky, ~2x change isn't a lot. Can you tell the exact model and capacity of this Micron, and what controller was used in this test?

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-21 Thread Eric K. Miller
We were able to isolate an individual Micron 5200 and perform Vitaliy's tests in his spreadsheet. An interesting item - write cache changes do NOT require a power cycle to take effect, at least on a Micron 5200. The complete results from fio are included at the end of this message for the

Re: [ceph-users] OSD crash after change of osd_memory_target

2020-01-21 Thread Stefan Kooman
Quoting Martin Mlynář (nexus+c...@smoula.net): > Do you think this could help? OSD does not even start, I'm getting a little > lost how flushing caches could help. I might have mis-understood. I though the OSDs crashed when you set the config setting. > According to trace I suspect something

Re: [ceph-users] OSD crash after change of osd_memory_target

2020-01-21 Thread Martin Mlynář
Dne út 21. 1. 2020 17:09 uživatel Stefan Kooman napsal: > Quoting Martin Mlynář (nexus+c...@smoula.net): > > > > > When I remove this option: > > # ceph config rm osd osd_memory_target > > > > OSD starts without any trouble. I've seen same behaviour when I wrote > > this parameter into

Re: [ceph-users] CephFS with cache-tier kernel-mount client unable to write (Nautilus)

2020-01-21 Thread Ilya Dryomov
On Tue, Jan 21, 2020 at 7:51 PM Hayashida, Mami wrote: > > Ilya, > > Thank you for your suggestions! > > `dmsg` (on the client node) only had `libceph: mon0 10.33.70.222:6789 socket > error on write`. No further detail. But using the admin key (client.admin) > for mounting CephFS solved my

Re: [ceph-users] S3 Bucket usage up 150% diference between rgw-admin and external metering tools.

2020-01-21 Thread EDH - Manuel Rios
Hi Robin, - What are the external tools? CloudBerry S3 Explorer and S3 Browser - How many objects do the external tools report as existing? Tool report 72142 keys (Aprox 6TB) vs CEPH num_objects 180981 (9TB) - Do the external tools include incomplete multipart uploads in their size data? I

Re: [ceph-users] CephFS with cache-tier kernel-mount client unable to write (Nautilus)

2020-01-21 Thread Hayashida, Mami
Ilya, Thank you for your suggestions! `dmsg` (on the client node) only had `libceph: mon0 10.33.70.222:6789 socket error on write`. No further detail. But using the admin key (client.admin) for mounting CephFS solved my problem. I was able to write successfully! :-) $ sudo mount -t ceph

Re: [ceph-users] CephFS with cache-tier kernel-mount client unable to write (Nautilus)

2020-01-21 Thread Ilya Dryomov
On Tue, Jan 21, 2020 at 6:02 PM Hayashida, Mami wrote: > > I am trying to set up a CephFS with a Cache Tier (for data) on a mini test > cluster, but a kernel-mount CephFS client is unable to write. Cache tier > setup alone seems to be working fine (I tested it with `rados put` and `osd > map`

Re: [ceph-users] S3 Bucket usage up 150% diference between rgw-admin and external metering tools.

2020-01-21 Thread Robin H. Johnson
On Mon, Jan 20, 2020 at 12:57:51PM +, EDH - Manuel Rios wrote: > Hi Cephs > > Several nodes of our Ceph 14.2.5 are fully dedicated to host cold storage / > backups information. > > Today checking the data usage with a customer found that rgw-admin is > reporting: ... > That's near 5TB used

[ceph-users] CephFS with cache-tier kernel-mount client unable to write (Nautilus)

2020-01-21 Thread Hayashida, Mami
I am trying to set up a CephFS with a Cache Tier (for data) on a mini test cluster, but a kernel-mount CephFS client is unable to write. Cache tier setup alone seems to be working fine (I tested it with `rados put` and `osd map` commands to verify on which OSDs the objects are placed) and setting

[ceph-users] MDS: obscene buffer_anon memory use when scanning lots of files

2020-01-21 Thread John Madden
On 14.2.5 but also present in Luminous, buffer_anon memory use spirals out of control when scanning many thousands of files. The use case is more or less "look up this file and if it exists append this chunk to it, otherwise create it with this chunk." The memory is recovered as soon as the

Re: [ceph-users] OSD crash after change of osd_memory_target

2020-01-21 Thread Stefan Kooman
Quoting Martin Mlynář (nexus+c...@smoula.net): > > When I remove this option: > # ceph config rm osd osd_memory_target > > OSD starts without any trouble. I've seen same behaviour when I wrote > this parameter into /etc/ceph/ceph.conf > > Is this a known bug? Am I doing something wrong? I

[ceph-users] OSD crash after change of osd_memory_target

2020-01-21 Thread Martin Mlynář
Hi, I'm having troubles changing osd_memory_target on my test cluster. I've upgraded whole cluster from luminous to nautiuls, all OSDs are running bluestore. Because this testlab is short in RAM, I wanted to lower osd_memory_target to save some memory. # ceph version ceph version 14.2.6

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-21 Thread Frank Schilder
OK, now I understand. Yes, the cache setting will take effect immediately. Its more about do you trust the disk firmware to apply the change correctly in all situations when production IO is active at the same time (will volatile cache be flushed correctly or not)? I would not and rather change

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-21 Thread Stefan Priebe - Profihost AG
Hello Igor, thanks for all your feedback and all your help. The first thing i'll try is to upgrade a bunch of system from 4.19.66 kernel to 4.19.97 and see what happens. I'll report back in 7-10 days to verify whether this helps. Greets, Stefan Am 20.01.20 um 13:12 schrieb Igor Fedotov: > Hi

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-21 Thread Sasha Litvak
Frank, Sorry for the confusion. I thought that turning off cache using hdparm -W 0 /dev/sdx takes effect right away and in case of non-raid controllers and Seagate or Micron SSDs I would see a difference starting fio benchmark right after executing hdparm. So I wonder it makes a difference

[ceph-users] Understand ceph df details

2020-01-21 Thread CUZA Frédéric
Hi everyone, I'm trying to understand where is the difference between the command : ceph df details And the result I'm getting when I run this script : total_bytes=0 while read user; do echo $user bytes=$(radosgw-admin user stats --uid=${user} | grep total_bytes_rounded | tr -dc "0-9") if

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-21 Thread Frank Schilder
> So hdparam -W 0 /dev/sdx doesn't work or it makes no difference? I wrote "We found the raw throughput in fio benchmarks to be very different for write-cache enabled and disabled, exactly as explained in the performance article.", so yes, it makes a huge difference. > Also I am not sure I

Re: [ceph-users] Ceph MDS randomly hangs with no useful error message

2020-01-20 Thread Yan, Zheng
On Tue, Jan 21, 2020 at 12:09 AM Janek Bevendorff wrote: > > Hi, I did as you asked and created a thread dump with GDB on the > blocking MDS. Here's the result: https://pastebin.com/pPbNvfdb > I don't find any clue from the backtrace. please run 'ceph daemon mds. dump_historic_ops' and

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-20 Thread Eric K. Miller
Hi Vitaliy, > You say you don't have access to raw drives. What does it mean? Do you > run Ceph OSDs inside VMs? In that case you should probably disable > Micron caches on the hosts, not just in VMs. Sorry, I should have been more clear. This cluster is in production, so I needed to schedule

[ceph-users] lists and gmail

2020-01-20 Thread Sasha Litvak
It seems that people now split between new and old list servers. Regardless of either one of them, I am missing a number of messages that appear on archive pages but never seem to make to my inbox. And no they are not in my junk folder. I wonder if some of my questions are not getting a

Re: [ceph-users] Ceph MDS randomly hangs with no useful error message

2020-01-20 Thread Janek Bevendorff
Hi, I did as you asked and created a thread dump with GDB on the blocking MDS. Here's the result: https://pastebin.com/pPbNvfdb On 17/01/2020 13:07, Yan, Zheng wrote: On Fri, Jan 17, 2020 at 4:47 PM Janek Bevendorff wrote: Hi, We have a CephFS in our cluster with 3 MDS to which > 300

Re: [ceph-users] CephsFS client hangs if one of mount-used MDS goes offline

2020-01-20 Thread Anton Aleksandrov
Yes, another MDS takes it over and even comes back, but client does not always "unfreeze". Weird, i see some different versions.. ceph versions {     "mon": {     "ceph version 13.2.6 (7b695f835b03642f85998b2ae7b6dd093d9fbce4) mimic (stable)": 2,     "ceph version 13.2.8

Re: [ceph-users] CephsFS client hangs if one of mount-used MDS goes offline

2020-01-20 Thread Wido den Hollander
On 1/20/20 4:17 PM, Anton Aleksandrov wrote: > Hello community, > > We have very small ceph cluster of just 12 OSDs (1 per small server), 3 > MDS (one active) and 1 cephFS client. > Which version of Ceph? $ ceph versions > CephFS client is running Centos7, kernel

[ceph-users] CephsFS client hangs if one of mount-used MDS goes offline

2020-01-20 Thread Anton Aleksandrov
Hello community, We have very small ceph cluster of just 12 OSDs (1 per small server), 3 MDS (one active) and 1 cephFS client. CephFS client is running Centos7, kernel 3.10.0-957.27.2.el7.x86_64. We created 3 MDS servers for redundancy and we mount our filesystem by connecting to 3 of them.

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-20 Thread vitalif
Hi Eric, You say you don't have access to raw drives. What does it mean? Do you run Ceph OSDs inside VMs? In that case you should probably disable Micron caches on the hosts, not just in VMs. Yes, disabling the write cache only takes place upon a power cycle... or upon the next hotplug of

[ceph-users] ceph 14.2.6 problem with default args to rbd (--name)

2020-01-20 Thread Rainer Krienke
Hello, I am fighting with rbd and CEPH_ARGS in order to make typing easier on a client. First I created a keyring on one of the ceph nodes: # ceph auth add client.rainer mon 'profile rbd' osd 'profile rbd' added key for client.rainer Then I added this keyring to /etc/ceph/ceph.keyring on a

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-20 Thread Sasha Litvak
So hdparam -W 0 /dev/sdx doesn't work or it makes no difference? Also I am not sure I understand why it should happen before OSD have been started. At least in my experience hdparam does it to hardware regardless. On Mon, Jan 20, 2020, 2:25 AM Frank Schilder wrote: > We are using Micron 5200

[ceph-users] S3 Bucket usage up 150% diference between rgw-admin and external metering tools.

2020-01-20 Thread EDH - Manuel Rios
Hi Cephs Several nodes of our Ceph 14.2.5 are fully dedicated to host cold storage / backups information. Today checking the data usage with a customer found that rgw-admin is reporting: { "bucket": "XX", "tenant": "", "zonegroup": "4d8c7c5f-ca40-4ee3-b5bb-b2cad90bd007",

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-20 Thread Igor Fedotov
Hi Stefan, these lines are result of transaction dump performed on a failure during transaction submission (which is shown as "submit_transaction error: Corruption: block checksum mismatch code = 2" Most probably they are out of interest (checksum errors are unlikely to be caused by

Re: [ceph-users] OSD up takes 15 minutes after machine restarts

2020-01-20 Thread Igor Fedotov
No, bluestore_fsck_on_mount_deep is applied when bluestore_fsck_on_mount is set to true only. Hence there is no fsck on mount in your case. Thanks, Igor On 1/20/2020 10:25 AM, huxia...@horebdata.cn wrote: HI, Igor, does this could cause the problem?

Re: [ceph-users] Issues with Nautilus 14.2.6 ceph-volume lvm batch --bluestore ?

2020-01-20 Thread Janne Johansson
Den mån 20 jan. 2020 kl 09:03 skrev Dave Hall : > Hello, > Since upgrading to Nautilus (+ Debian 10 Backports), when I issue > 'ceph-volume lvm batch --bluestore ' it fails with > > bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid > > I previously had Luminous + Debian 9

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-20 Thread Frank Schilder
We are using Micron 5200 PRO, 1.92TB for RBD images on KVM and are very happy with the performance. We are using EC 6+2 pools, which really eat up IOPs. Still, we get enough performance out to run 20-50 VMs per disk, which results in good space utilisation as well since our default image size

Re: [ceph-users] OSD up takes 15 minutes after machine restarts

2020-01-19 Thread huxia...@horebdata.cn
HI, Igor, does this could cause the problem? huxia...@horebdata.cn From: Igor Fedotov Date: 2020-01-19 11:41 To: huxia...@horebdata.cn; ceph-users Subject: Re: [ceph-users] OSD up takes 15 minutes after machine restarts Hi Samuel, wondering if you have bluestore_fsck_on_mount option set to

Re: [ceph-users] Issues with Nautilus 14.2.6 ceph-volume lvm batch --bluestore ?

2020-01-19 Thread Dave Hall
Nigel, Thanks.  I've never seen that.  Cool. -Dave Dave Hall Binghamton University On 1/19/2020 11:15 PM, Nigel Williams wrote: On Mon, 20 Jan 2020 at 14:15, Dave Hall wrote: BTW, I did try to search the list archives via http://lists.ceph.com/pipermail/ceph-users-ceph.com/, but that

Re: [ceph-users] Issues with Nautilus 14.2.6 ceph-volume lvm batch --bluestore ?

2020-01-19 Thread Nigel Williams
On Mon, 20 Jan 2020 at 14:15, Dave Hall wrote: > BTW, I did try to search the list archives via > http://lists.ceph.com/pipermail/ceph-users-ceph.com/, but that didn't work > well for me. Is there another way to search? With your favorite search engine (say Goog / ddg ), you can do this:

[ceph-users] Issues with Nautilus 14.2.6 ceph-volume lvm batch --bluestore ?

2020-01-19 Thread Dave Hall
Hello, Since upgrading to Nautilus (+ Debian 10 Backports), when I issue 'ceph-volume lvm batch --bluestore ' it fails with bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid I previously had Luminous + Debian 9 running on the same hardware with the same OSD layout, but I

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-19 Thread Stefan Priebe - Profihost AG
Hello Igor, there's absolutely nothing in the logs before. What do those lines mean: Put( Prefix = O key = 0x7f8001cc45c881217262'd_data.4303206b8b4567.9632!='0xfffe6f0012'x' Value size = 480) Put( Prefix = O key =

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-19 Thread Stefan Priebe - Profihost AG
Yes, except that this happens on 8 different clusters with different hw but same ceph version and same kernel version. Greets, Stefan > Am 19.01.2020 um 11:53 schrieb Igor Fedotov : > > So the intermediate summary is: > > Any OSD in the cluster can experience interim RocksDB checksum

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-19 Thread Igor Fedotov
So the intermediate summary is: Any OSD in the cluster can experience interim RocksDB checksum failure. Which isn't present after OSD restart. No HW issues observed, no persistent artifacts (except OSD log) afterwards. And looks like the issue is rather specific to the cluster as no similar

Re: [ceph-users] OSD up takes 15 minutes after machine restarts

2020-01-19 Thread Igor Fedotov
Hi Samuel, wondering if you have bluestore_fsck_on_mount option set to true? Can you see high read load over OSD device(s) during the startup? If so it might be fsck running which takes that long. Thanks, Igor On 1/19/2020 11:53 AM, huxia...@horebdata.cn wrote: Dear folks, I had a

[ceph-users] [ceph-osd ] osd can not boot

2020-01-19 Thread Wei Zhao
Hi : A server was just rebooted and the osd cant boot .The log is the following. -3> 2020-01-19 17:39:25.904673 7f5b8e5e9d80 -1 bluestore(/var/lib/ceph/osd/ceph-44) _verify_csum bad crc32c/0x1000 checksum at blob offset 0x0, got 0xd2acc81f, expected 0x62cf539d, device location

[ceph-users] OSD up takes 15 minutes after machine restarts

2020-01-19 Thread huxia...@horebdata.cn
Dear folks, I had a strange situation with 3-node Ceph cluster on Luminous 12.2.12 with bluestore. Each machine has 5 OSDs on HDD, and each OSD uses a 30GB DB/WAL partition on SSD. At the beginning without much data, OSDs can quickly up if one node restarts. Then I ran 4-day long stress

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-18 Thread Eric K. Miller
Hi Vitaliy, Similar to Stefan, we have a bunch of Micron 5200's (3.84TB ECO SATA version) in a Ceph cluster (Nautilus) and performance seems less than optimal. I have followed all instructions on your site (thank you for your wonderful article btw!!), but I haven't seen much change.

Re: [ceph-users] Monitor handle_auth_bad_method

2020-01-18 Thread Justin Engwer
Gatherkeys and config push seem to have done the job. Thanks for your help, Paul! Justin On Sat., Jan. 18, 2020, 02:33 Paul Emmerich, wrote: > check if the mons have the same keyring file and the same config file. > -- > Paul Emmerich > > Looking for help with your Ceph cluster? Contact us at

Re: [ceph-users] Monitor handle_auth_bad_method

2020-01-18 Thread Paul Emmerich
check if the mons have the same keyring file and the same config file. -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Sat, Jan 18, 2020 at 12:39 AM Justin Engwer wrote:

Re: [ceph-users] Default Pools

2020-01-18 Thread Paul Emmerich
RGW tools will automatically deploy these pools, for example, running radosgw-admin will create them if they don't exist. Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585

Re: [ceph-users] Slow Performance - Sequential IO

2020-01-17 Thread Christian Balzer
Hello, I had very odd results in the past with the fio rbd engine and would suggest testing things in the environment you're going to deploy in, end to end. That said, without any caching and coalescing of writes, sequential 4k writes will hit the same set of OSDs for 4MB worth of data, thus

Re: [ceph-users] Default Pools

2020-01-17 Thread Daniele Riccucci
Hello, I'm still a bit confused by the .rgw.root and the default.rgw.{control,meta,log} pools. I recently removed the RGW daemon I had running and the aforementioned pools, however after a rebalance I suddenly find them again in the output of: $ ceph osd pool ls cephfs_data cephfs_metadata

[ceph-users] Monitor handle_auth_bad_method

2020-01-17 Thread Justin Engwer
Hi, I'm a home user of ceph. Most of the time I can look at the email lists and articles and figure things out on my own. I've unfortunately run into an issue I can't troubleshoot myself. Starting one of my monitors yields this error: 2020-01-17 15:34:13.497 7fca3d006040 0 mon.kvm2@-1(probing)

Re: [ceph-users] Slow Performance - Sequential IO

2020-01-17 Thread Anthony Brandelli (abrandel)
Not been able to make any headway on this after some significant effort. -Tested all 48 SSDs with FIO directly, all tested with 10% of each other for 4k iops in rand|seq read|write. -Disabled all CPU power save. -Tested with both rbd cache enabled and disabled on the client. -Tested with drive

Re: [ceph-users] Beginner questions

2020-01-17 Thread Dave Hall
Frank, Thank you for your input.  It is good to know that the cluster will go read-only in if a node goes down.  Our circumstance is probably a bit unusual, which is why I'm considering the2+1 solution.  We have a researcher who will be collecting extremely large amounts of data in real

Re: [ceph-users] Weird mount issue (Ubuntu 18.04, Ceph 14.2.5 & 14.2.6)

2020-01-17 Thread Jeff Layton
On Fri, 2020-01-17 at 17:10 +0100, Ilya Dryomov wrote: > On Fri, Jan 17, 2020 at 2:21 AM Aaron wrote: > > No worries, can definitely do that. > > > > Cheers > > Aaron > > > > On Thu, Jan 16, 2020 at 8:08 PM Jeff Layton wrote: > > > On Thu, 2020-01-16 at 18:42 -0500, Jeff Layton wrote: > > > >

Re: [ceph-users] Weird mount issue (Ubuntu 18.04, Ceph 14.2.5 & 14.2.6)

2020-01-17 Thread Ilya Dryomov
On Fri, Jan 17, 2020 at 2:21 AM Aaron wrote: > > No worries, can definitely do that. > > Cheers > Aaron > > On Thu, Jan 16, 2020 at 8:08 PM Jeff Layton wrote: >> >> On Thu, 2020-01-16 at 18:42 -0500, Jeff Layton wrote: >> > On Wed, 2020-01-15 at 08:05 -0500, Aaron wrote: >> > > Seeing a weird

Re: [ceph-users] Weird mount issue (Ubuntu 18.04, Ceph 14.2.5 & 14.2.6)

2020-01-17 Thread Jeff Layton
Actually, scratch that. I went ahead and opened this: https://tracker.ceph.com/issues/43649 Feel free to watch that one for updates. On Fri, 2020-01-17 at 07:43 -0500, Jeff Layton wrote: > No problem. Can you let me know the tracker bug number once you've > opened it? > > Thanks, > Jeff > >

Re: [ceph-users] Ceph MDS randomly hangs with no useful error message

2020-01-17 Thread Janek Bevendorff
Thanks. I will do that. Right now, we have quite a few lags when listing folders, which is probably due to another client heavily using the system. Unfortunately, it's rather hard to debug at the moment, since the suspected client has to use our Ganesha bridge instead of connecting to the Ceph

Re: [ceph-users] Ceph MDS randomly hangs with no useful error message

2020-01-17 Thread Yan, Zheng
On Fri, Jan 17, 2020 at 4:47 PM Janek Bevendorff wrote: > > Hi, > > We have a CephFS in our cluster with 3 MDS to which > 300 clients > connect at any given time. The FS contains about 80 TB of data and many > million files, so it is important that meta data operations work > smoothly even when

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-17 Thread Stefan Priebe - Profihost AG
HI Igor, Am 17.01.20 um 12:10 schrieb Igor Fedotov: > hmmm.. > > Just in case - suggest to check H/W errors with dmesg. this happens on around 80 nodes - i don't expect all of those have not identified hw errors. Also all of them are monitored - no dmesg outpout contains any errors. > Also

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-17 Thread Igor Fedotov
hmmm.. Just in case - suggest to check H/W errors with dmesg. Also there are some (not very much though) chances this is another incarnation of the following bug: https://tracker.ceph.com/issues/22464 https://github.com/ceph/ceph/pull/24649 The corresponding PR works around it for main

Re: [ceph-users] Beginner questions

2020-01-17 Thread Frank Schilder
I would strongly advise against 2+1 EC pools for production if stability is your main concern. There was a discussion towards the end of last year addressing this in more detail. Short story, if you don't have at least 8-10 nodes (in the short run), EC is not suitable. You cannot maintain a

[ceph-users] Ceph MDS randomly hangs with no useful error message

2020-01-17 Thread Janek Bevendorff
Hi, We have a CephFS in our cluster with 3 MDS to which > 300 clients connect at any given time. The FS contains about 80 TB of data and many million files, so it is important that meta data operations work smoothly even when listing large directories. Previously, we had massive stability

Re: [ceph-users] Beginner questions

2020-01-16 Thread Bastiaan Visser
There is no difference in allocation between replication or EC. If failure domain is host, one osd per host ok s used for a PG. So if you use a 2+1 EC profile with a host failure domain, you need 3 hosts for a healthy cluster. The pool will go read-only when you have a failure (host or disk), or

Re: [ceph-users] Beginner questions

2020-01-16 Thread Dave Hall
Bastiaan, Regarding EC pools:   Our concern at 3 nodes is that 2-way replication seems risky - if the two copies don't match, which one is corrupted.  However,  3-way replication on a 3 node cluster triples the price per TB.   Doing EC pools that are the equivalent of RAID-5 2+1 seems like

Re: [ceph-users] Weird mount issue (Ubuntu 18.04, Ceph 14.2.5 & 14.2.6)

2020-01-16 Thread Aaron
This debugging started because the ceph-provisioner from k8s was making those users...but what we found was doing something similar by hand caused the same issue. Just surprised no one else using k8s and ceph backed PVC/PVs ran into this issue. Thanks again for all your help! Cheers Aaron On

Re: [ceph-users] Weird mount issue (Ubuntu 18.04, Ceph 14.2.5 & 14.2.6)

2020-01-16 Thread Aaron
No worries, can definitely do that. Cheers Aaron On Thu, Jan 16, 2020 at 8:08 PM Jeff Layton wrote: > On Thu, 2020-01-16 at 18:42 -0500, Jeff Layton wrote: > > On Wed, 2020-01-15 at 08:05 -0500, Aaron wrote: > > > Seeing a weird mount issue. Some info: > > > > > > No LSB modules are

Re: [ceph-users] [External Email] RE: Beginner questions

2020-01-16 Thread DHilsbos
Paul; So is the 3/30/300GB a limit of RocksDB, or of Bluestore? The percentages you list, are they used DB / used data? If so... Where do you get the used DB data from? Thank you, Dominic L. Hilsbos, MBA Director – Information Technology Perform Air International Inc.

Re: [ceph-users] [External Email] RE: Beginner questions

2020-01-16 Thread Paul Emmerich
Discussing DB size requirements without knowing the exact cluster requirements doesn't work. Here are some real-world examples: cluster1: CephFS, mostly large files, replicated x3 0.2% used for metadata cluster2: radosgw, mix between replicated and erasure, mixed file sizes (lots of tiny files,

[ceph-users] Snapshots and Backup from Horizon to ceph s3 buckets

2020-01-16 Thread Radhakrishnan2 S
Hello, We are trying to route backups & snapshots of cinder volume and nova instances, into the s3 buckets hosted on ceph. Currently ceph is the block storage target as well. What we want to achieve ? 1. all snapshots of cinder volumes / nova instance to be routed to s3 buckets of that

Re: [ceph-users] [External Email] RE: Beginner questions

2020-01-16 Thread Bastiaan Visser
Dave made a good point WAL + DB might end up a little over 60G, I would probably go with ~70Gig partitions /LV's per OSD in your case. (if the nvme drive is smart enough to spread the writes over all available capacity, mort recent nvme's are). I have not yet seen a WAL larger or even close to

Re: [ceph-users] ceph nautilus cluster name

2020-01-16 Thread Ignazio Cassano
Hello Stefan, but if I want to use rbd mirroring I must have site-a.conf and site-b.conf on one of my nodes.probably one of the mon nodes. Is it only a configuration on ceph client side ? Thanks Ignazio Il Gio 16 Gen 2020, 22:13 Stefan Kooman ha scritto: > Quoting Ignazio Cassano

Re: [ceph-users] ceph nautilus cluster name

2020-01-16 Thread Stefan Kooman
Quoting Ignazio Cassano (ignaziocass...@gmail.com): > Hello, I just deployed nautilus with ceph-deploy. > I did not find any option to give a cluster name to my ceph so its name is > "ceph". > Please, how can I chenge my cluster name without reinstalling ? > > Please, how can I set the cluster

[ceph-users] ceph nautilus cluster name

2020-01-16 Thread Ignazio Cassano
Hello, I just deployed nautilus with ceph-deploy. I did not find any option to give a cluster name to my ceph so its name is "ceph". Please, how can I chenge my cluster name without reinstalling ? Please, how can I set the cluster name in installation phase ? Many thanks for help Ignazio

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Stefan Priebe - Profihost AG
Hi Igor, answers inline. Am 16.01.20 um 21:34 schrieb Igor Fedotov: > you may want to run fsck against failing OSDs. Hopefully it will shed > some light. fsck just says everything fine: # ceph-bluestore-tool --command fsck --path /var/lib/ceph/osd/ceph-27/ fsck success > Also wondering if

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Igor Fedotov
Stefan, you may want to run fsck against failing OSDs. Hopefully it will shed some light. Also wondering if OSD is able to recover (startup and proceed working) after facing the issue? If so do you have any one which failed multiple times? Do you have logs for these occurrences? Also

Re: [ceph-users] [External Email] RE: Beginner questions

2020-01-16 Thread DHilsbos
Dave; I don't like reading inline responses, so... I have zero experience with EC pools, so I won't pretend to give advice in that area. I would think that small NVMe for DB would be better than nothing, but I don't know. Once I got the hang of building clusters, it was relatively easy to

Re: [ceph-users] [External Email] RE: Beginner questions

2020-01-16 Thread Dave Hall
Dominic, We ended up with a 1.6TB PCIe NVMe in each node.  For 8 drives this worked out to a DB size of something like 163GB per OSD. Allowing for expansion to 12 drives brings it down to 124GB. So maybe just put the WALs on NVMe and leave the DBs on the platters? Understood that we will

Re: [ceph-users] Beginner questions

2020-01-16 Thread DHilsbos
Dave; I'd like to expand on this answer, briefly... The information in the docs is wrong. There have been many discussions about changing it, but no good alternative has been suggested, thus it hasn't been changed. The 3rd party project that Ceph's BlueStore uses for its database (RocksDB),

Re: [ceph-users] [External Email] Re: Beginner questions

2020-01-16 Thread Dave Hall
Paul, Bastiaan, Thank you for your responses and for alleviating my concerns about Nautilus.  The good news is that I can still easily move up to Debian 10.  BTW, I assume that this is still with the 4.19 kernel? Also, I'd like to inject additional customizations into my Debian configs via

Re: [ceph-users] Beginner questions

2020-01-16 Thread Paul Emmerich
Don't use Mimic, support for it is far worse than Nautilus or Luminous. I think we were the only company who built a product around Mimic, both Redhat and Suse enterprise storage was Luminous and then Nautilus skipping Mimic entirely. We only offered Mimic as a default for a limited time and

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Stefan Priebe - Profihost AG
Hi Igor, ouch sorry. Here we go: -1> 2020-01-16 01:10:13.404090 7f3350a14700 -1 rocksdb: submit_transaction error: Corruption: block checksum mismatch code = 2 Rocksdb transaction: Put( Prefix = M key =

Re: [ceph-users] Beginner questions

2020-01-16 Thread Bastiaan Visser
I would definitely go for Nautilus. there are quite some optimizations that went in after mimic. Bluestore DB size usually ends up at either 30 or 60 GB. 30 GB is one of the sweet spots during normal operation. But during compaction, ceph writes the new data before removing the old, hence the

[ceph-users] Beginner questions

2020-01-16 Thread Dave Hall
Hello all. Sorry for the beginner questions... I am in the process of setting up a small (3 nodes, 288TB) Ceph cluster to store some research data.  It is expected that this cluster will grow significantly in the next year, possibly to multiple petabytes and 10s of nodes.  At this time I'm

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Igor Fedotov
Hi Stefan, would you please share log snippet prior the assertions? Looks like RocksDB is failing during transaction submission... Thanks, Igor On 1/16/2020 11:56 AM, Stefan Priebe - Profihost AG wrote: Hello, does anybody know a fix for this ASSERT / crash? 2020-01-16 02:02:31.316394

[ceph-users] Snapshots and Backup from Horizon to ceph s3 buckets

2020-01-16 Thread Radhakrishnan2 S
Hello, We are trying to route backups & snapshots of cinder volume and nova instances, into the s3 buckets hosted on ceph. Currently ceph is the block storage target as well. What we want to achieve ? 1. all snapshots of cinder volumes / nova instance to be routed to s3 buckets of that

  1   2   3   4   5   6   7   8   9   10   >