Re: [ceph-users] default.rgw.log contains large omap object

2019-10-14 Thread Paul Emmerich
Yeah, the number of shards is configurable ("rgw usage num shards"? or something). Are you sure you aren't using it? This feature is not enabled by default, someone had to explicitly set "rgw enable usage log" for you to run into this problem. Paul -- Paul Emmerich Looking for help with your

Re: [ceph-users] default.rgw.log contains large omap object

2019-10-14 Thread Troy Ablan
Paul, Apparently never. Appears to (potentially) have every request from the beginning of time (late last year, in my case). In our use case, we don't really need this data (not multi-tenant), so I might simply clear it. But in the case where this were an extremely high transaction cluster

Re: [ceph-users] default.rgw.log contains large omap object

2019-10-14 Thread Paul Emmerich
Looks like the usage log (radosgw-admin usage show), how often do you trim it? -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Mon, Oct 14, 2019 at 11:55 PM Troy Ablan

[ceph-users] default.rgw.log contains large omap object

2019-10-14 Thread Troy Ablan
Hi folks, Mimic cluster here, RGW pool with only default zone. I have a persistent error here LARGE_OMAP_OBJECTS 1 large omap objects 1 large objects found in pool 'default.rgw.log' Search the cluster log for 'Large omap object found' for more details. I think I've narrowed it

Re: [ceph-users] problem returning mon back to cluster

2019-10-14 Thread Paul Emmerich
How big is the mon's DB? As in just the total size of the directory you copied FWIW I recently had to perform mon surgery on a 14.2.4 (or was it 14.2.2?) cluster with 8 GB mon size and I encountered no such problems while syncing a new mon which took 10 minutes or so. Paul -- Paul Emmerich

Re: [ceph-users] problem returning mon back to cluster

2019-10-14 Thread Nikola Ciprich
On Mon, Oct 14, 2019 at 04:31:22PM +0200, Nikola Ciprich wrote: > On Mon, Oct 14, 2019 at 01:40:19PM +0200, Harald Staub wrote: > > Probably same problem here. When I try to add another MON, "ceph > > health" becomes mostly unresponsive. One of the existing ceph-mon > > processes uses 100% CPU for

Re: [ceph-users] Ceph Negative Objects Number

2019-10-14 Thread Lazuardi Nasution
Hi Igor, It is the old Jewel (v10.2.11). This case is happen after I do cache-try-flush-evict-all or cache-flush-evict-all on respected tier pool. Best regards, On Mon, Oct 14, 2019 at 7:38 PM Igor Fedotov wrote: > Hi Lazuardi, > > never seen that. Just wondering what Ceph version are you

Re: [ceph-users] problem returning mon back to cluster

2019-10-14 Thread Nikola Ciprich
On Mon, Oct 14, 2019 at 01:40:19PM +0200, Harald Staub wrote: > Probably same problem here. When I try to add another MON, "ceph > health" becomes mostly unresponsive. One of the existing ceph-mon > processes uses 100% CPU for several minutes. Tried it on 2 test > clusters (14.2.4, 3 MONs, 5

Re: [ceph-users] Openstack VM IOPS drops dramatically during Ceph recovery

2019-10-14 Thread Robert LeBlanc
On Thu, Oct 10, 2019 at 2:23 PM huxia...@horebdata.cn wrote: > > Hi, folks, > > I have a middle-size Ceph cluster as cinder backup for openstack (queens). > Duing testing, one Ceph node went down unexpected and powered up again ca 10 > minutes later, Ceph cluster starts PG recovery. To my

Re: [ceph-users] tcmu-runner: mismatched sizes for rbd image size

2019-10-14 Thread Kilian Ries
@Mike Did you have the chance to update download.ceph.com repositories for the new version? I just tested the packages from shaman in our DEV environment and it seems to fix the work - after updating the packages i was not able to reproduce the error again and tcmu-runner starts up without

Re: [ceph-users] Ceph Negative Objects Number

2019-10-14 Thread Igor Fedotov
Hi Lazuardi, never seen that. Just wondering what Ceph version are you running? Thanks, Igor On 10/8/2019 3:52 PM, Lazuardi Nasution wrote: Hi, I get following weird negative objects number on tiering. Why is this happening? How to get back to normal? Best regards, [root@management-a

Re: [ceph-users] Ceph Negative Objects Number

2019-10-14 Thread Lazuardi Nasution
Hi, Any body has same problem with my case? Best regards, On Tue, Oct 8, 2019, 19:52 Lazuardi Nasution wrote: > Hi, > > I get following weird negative objects number on tiering. Why is this > happening? How to get back to normal? > > Best regards, > > [root@management-a ~]# ceph df detail >

Re: [ceph-users] problem returning mon back to cluster

2019-10-14 Thread Harald Staub
Probably same problem here. When I try to add another MON, "ceph health" becomes mostly unresponsive. One of the existing ceph-mon processes uses 100% CPU for several minutes. Tried it on 2 test clusters (14.2.4, 3 MONs, 5 storage nodes with around 2 hdd osds each). To avoid errors like "lease

Re: [ceph-users] Pool statistics via API

2019-10-14 Thread Sinan Polat
Hi Ernesto, I just opened the Dashboard and there is no menu at the top-right. Also no "?". I have a menu at the top-left which has the following items: Cluster health, Cluster, Block and Filesystems. Running Ceph version 12.2.8-89. Kind regards, Sinan Polat > Op 11 oktober 2019 om 22:09

[ceph-users] problem returning mon back to cluster

2019-10-13 Thread Nikola Ciprich
dear ceph users and developers, on one of our production clusters, we got into pretty unpleasant situation. After rebooting one of the nodes, when trying to start monitor, whole cluster seems to hang, including IO, ceph -s etc. When this mon is stopped again, everything seems to continue.

Re: [ceph-users] mds fail ing to start 14.2.2

2019-10-11 Thread Yan, Zheng
On Sat, Oct 12, 2019 at 1:10 AM Kenneth Waegeman wrote: > Hi all, > > After solving some pg inconsistency problems, my fs is still in > trouble. my mds's are crashing with this error: > > > > -5> 2019-10-11 19:02:55.375 7f2d39f10700 1 mds.1.564276 rejoin_start > > -4> 2019-10-11

Re: [ceph-users] mds servers in endless segfault loop

2019-10-11 Thread Pickett, Neale T
I have created an anonymized crash log at https://pastebin.ubuntu.com/p/YsVXQQTBCM/ in the hopes that it can help someone understand what's leading to our MDS outage. Thanks in advance for any assistance. From: Pickett, Neale T Sent: Thursday, October 10,

Re: [ceph-users] Pool statistics via API

2019-10-11 Thread Sinan Polat
Hi Ernesto, Thanks for the information! I didn’t know about the existence of the REST Dashboard API. I will check that out, Thanks again! Sinan > Op 11 okt. 2019 om 21:06 heeft Ernesto Puerta het > volgende geschreven: > > Hi Sinan, > > If it's in the Dashboard, it sure comes from the

Re: [ceph-users] Pool statistics via API

2019-10-11 Thread Ernesto Puerta
Hi Sinan, If it's in the Dashboard, it sure comes from the Dashboard REST API (which is an API completely unrelated to the RESTful Module). To check the Dashboard REST API, log in there and click on the top-right "?" menu, and in the dropdown, click on "API". That will lead you to the

[ceph-users] mds fail ing to start 14.2.2

2019-10-11 Thread Kenneth Waegeman
Hi all, After solving some pg inconsistency problems, my fs is still in trouble.  my mds's are crashing with this error:     -5> 2019-10-11 19:02:55.375 7f2d39f10700  1 mds.1.564276 rejoin_start     -4> 2019-10-11 19:02:55.385 7f2d3d717700  5 mds.beacon.mds01 received beacon reply

Re: [ceph-users] lot of inconsistent+failed_repair - failed to pick suitable auth object (14.2.3)

2019-10-11 Thread Kenneth Waegeman
On 11/10/2019 01:21, Brad Hubbard wrote: On Fri, Oct 11, 2019 at 12:27 AM Kenneth Waegeman wrote: Hi Brad, all, Pool 6 has min_size 2: pool 6 'metadata' replicated size 3 min_size 2 crush_rule 1 object_hash rjenkins pg_num 1024 pgp_num 1024 autoscale_mode warn last_change 172476 flags

Re: [ceph-users] ceph version 14.2.3-OSD fails

2019-10-11 Thread Stefan Priebe - Profihost AG
> Am 11.10.2019 um 14:07 schrieb Igor Fedotov : > >  > Hi! > > originally your issue looked like the ones from > https://tracker.ceph.com/issues/42223 > > And it looks like lack of some key information for FreeListManager in RocksDB. > > Once you have it present we can check the content of

Re: [ceph-users] ceph version 14.2.3-OSD fails

2019-10-11 Thread Igor Fedotov
Hi! originally your issue looked like the ones from https://tracker.ceph.com/issues/42223 And it looks like lack of some key information for FreeListManager in RocksDB. Once you have it present we can check the content of the RocksDB to prove this hypothesis, please let me know if you

Re: [ceph-users] rgw: multisite support

2019-10-11 Thread M Ranga Swami Reddy
I have setup the realm, zonegroup and master zone. Now Iam pulling the realm details from master to secondary...it failed with "request failed: (22) invalid argument" == radosgw-admin realm pull --url={url}, --access-key={key} --secret={key} request failed: (22) Invalid argument == On Mon, Oct

[ceph-users] mds servers in endless segfault loop

2019-10-10 Thread Pickett, Neale T
Hello, ceph-users. Our mds servers keep segfaulting from a failed assertion, and for the first time I can't find anyone else who's posted about this problem. None of them are able to stay up, so our cephfs is down. We recently had to truncate the journal log after an upgrade to nautilus, and

Re: [ceph-users] Pool statistics via API

2019-10-10 Thread Konstantin Shalygin
Currently I am getting the pool statistics (especially USED/MAX AVAIL) via the command line: ceph df -f json-pretty| jq '.pools[] | select(.name == "poolname") | .stats.max_avail' ceph df -f json-pretty| jq '.pools[] | select(.name == "poolname") | .stats.bytes_used' Command "ceph df" does not

Re: [ceph-users] lot of inconsistent+failed_repair - failed to pick suitable auth object (14.2.3)

2019-10-10 Thread Brad Hubbard
On Fri, Oct 11, 2019 at 12:27 AM Kenneth Waegeman wrote: > > Hi Brad, all, > > Pool 6 has min_size 2: > > pool 6 'metadata' replicated size 3 min_size 2 crush_rule 1 object_hash > rjenkins pg_num 1024 pgp_num 1024 autoscale_mode warn last_change 172476 > flags hashpspool stripe_width 0

[ceph-users] Openstack VM IOPS drops dramatically during Ceph recovery

2019-10-10 Thread huxia...@horebdata.cn
Hi, folks, I have a middle-size Ceph cluster as cinder backup for openstack (queens). Duing testing, one Ceph node went down unexpected and powered up again ca 10 minutes later, Ceph cluster starts PG recovery. To my surprise, VM IOPS drops dramatically during Ceph recovery, from ca. 13K IOPS

Re: [ceph-users] lot of inconsistent+failed_repair - failed to pick suitable auth object (14.2.3)

2019-10-10 Thread Kenneth Waegeman
Hi Brad, all, Pool 6 has min_size 2: pool 6 'metadata' replicated size 3 min_size 2 crush_rule 1 object_hash rjenkins pg_num 1024 pgp_num 1024 autoscale_mode warn last_change 172476 flags hashpspool stripe_width 0 application cephfs The output for all the inconsistent pgs is this: {    

Re: [ceph-users] radosgw pegging down 5 CPU cores when no data is being transferred

2019-10-10 Thread Paul Emmerich
I've also encountered this issue on a cluster yesterday; one CPU got stuck in an infinite loop in get_obj_data::flush and it stopped serving requests. I've updated the tracker issue accordingly. Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io

[ceph-users] Pool statistics via API

2019-10-10 Thread Sinan Polat
Hi, Currently I am getting the pool statistics (especially USED/MAX AVAIL) via the command line: ceph df -f json-pretty| jq '.pools[] | select(.name == "poolname") | .stats.max_avail' ceph df -f json-pretty| jq '.pools[] | select(.name == "poolname") | .stats.bytes_used' Command "ceph df" does

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-10 Thread Philippe D'Anjou
After trying to disable the paxos service trim temporarily (since that seemed to trigger it initially), we now see this:     "assert_condition": "from != to",     "assert_func": "void PaxosService::trim(MonitorDBStore::TransactionRef, version_t, version_t)",     "assert_file":

Re: [ceph-users] [Nfs-ganesha-devel] 2.7.3 with CEPH_FSAL Crashing

2019-10-10 Thread David C
Thanks, Patrick. Looks like the fix is awaiting review, I guess my options are to hold tight for 14.2.5 or patch myself if I get desperate. I've seen this crash about 4 times over the past 96 hours, is there anything I can do to mitigate the issue in the meantime? On Wed, Oct 9, 2019 at 9:23 PM

Re: [ceph-users] lot of inconsistent+failed_repair - failed to pick suitable auth object (14.2.3)

2019-10-10 Thread Brad Hubbard
Does pool 6 have min_size = 1 set? https://tracker.ceph.com/issues/24994#note-5 would possibly be helpful here, depending on what the output of the following command looks like. # rados list-inconsistent-obj [pgid] --format=json-pretty On Thu, Oct 10, 2019 at 8:16 PM Kenneth Waegeman wrote: >

[ceph-users] lot of inconsistent+failed_repair - failed to pick suitable auth object (14.2.3)

2019-10-10 Thread Kenneth Waegeman
Hi all, After some node failure and rebalancing, we have a lot of pg's in inconsistent state. I tried to repair, but it din't work. This is also in the logs: 2019-10-10 11:23:27.221 7ff54c9b0700  0 log_channel(cluster) log [DBG] : 6.327 repair starts 2019-10-10 11:23:27.431 7ff5509b8700 -1

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-10 Thread Philippe D'Anjou
How do I Import an osdmap in Nautilus? I saw documentation for older version but it seems one now can only export but not import anymore? Am Donnerstag, 10. Oktober 2019, 08:52:03 OESZ hat Philippe D'Anjou Folgendes geschrieben: I dont think this has anything to do with CephFS, the

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-09 Thread Philippe D'Anjou
I dont think this has anything to do with CephFS, the mon crashes for same reason even without the mds running.I have still the old rocksdb files but they had a corruption issue, not sure if that's easier to fix, there havent been any changes on the cluster in between. This is a disaster

Re: [ceph-users] Unexpected increase in the memory usage of OSDs

2019-10-09 Thread Gregory Farnum
On Wed, Oct 9, 2019 at 10:58 AM Vladimir Brik < vladimir.b...@icecube.wisc.edu> wrote: > Best I can tell, automatic cache sizing is enabled and all related > settings are at their default values. > > Looking through cache tunables, I came across > osd_memory_expected_fragmentation, which the docs

Re: [ceph-users] Ceph pg repair clone_missing?

2019-10-09 Thread Brad Hubbard
Awesome! Sorry it took so long. On Thu, Oct 10, 2019 at 12:44 AM Marc Roos wrote: > > > Brad, many thanks!!! My cluster has finally HEALTH_OK af 1,5 year or so! > :) > > > -Original Message- > Subject: Re: Ceph pg repair clone_missing? > > On Fri, Oct 4, 2019 at 6:09 PM Marc Roos >

Re: [ceph-users] [Nfs-ganesha-devel] 2.7.3 with CEPH_FSAL Crashing

2019-10-09 Thread Patrick Donnelly
Looks like this bug: https://tracker.ceph.com/issues/41148 On Wed, Oct 9, 2019 at 1:15 PM David C wrote: > > Hi Daniel > > Thanks for looking into this. I hadn't installed ceph-debuginfo, here's the > bt with line numbers: > > #0 operator uint64_t (this=0x10) at >

Re: [ceph-users] [Nfs-ganesha-devel] 2.7.3 with CEPH_FSAL Crashing

2019-10-09 Thread David C
Hi Daniel Thanks for looking into this. I hadn't installed ceph-debuginfo, here's the bt with line numbers: #0 operator uint64_t (this=0x10) at /usr/src/debug/ceph-14.2.2/src/include/object.h:123 #1 Client::fill_statx (this=this@entry=0x274b980, in=0x0, mask=mask@entry=341,

Re: [ceph-users] Unexpected increase in the memory usage of OSDs

2019-10-09 Thread Vladimir Brik
Best I can tell, automatic cache sizing is enabled and all related settings are at their default values. Looking through cache tunables, I came across osd_memory_expected_fragmentation, which the docs define as "estimate the percent of memory fragmentation". What's the formula to compute

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-09 Thread Gregory Farnum
On Mon, Oct 7, 2019 at 11:11 PM Philippe D'Anjou wrote: > > Hi, > unfortunately it's single mon, because we had major outage on this cluster > and it's just being used to copy off data now. We werent able to add more > mons because once a second mon was added it crashed the first one (there's a

Re: [ceph-users] Ceph multi site outage question

2019-10-09 Thread Melzer Pinto
Thanks - yeah jewel is old  But i meant to say nautilus and not luminous. The first option probably wont work for me. Since both sides are active and the application1 needs to write in both places as http://application1.something.com. The 2nd one in theory should work. I'm using haproxy and it

Re: [ceph-users] Unexpected increase in the memory usage of OSDs

2019-10-09 Thread Gregory Farnum
On Mon, Oct 7, 2019 at 7:20 AM Vladimir Brik wrote: > > > Do you have statistics on the size of the OSDMaps or count of them > > which were being maintained by the OSDs? > No, I don't think so. How can I find this information? Hmm I don't know if we directly expose the size of maps. There are

Re: [ceph-users] Ceph multi site outage question

2019-10-09 Thread Ed Fisher
Boy, Jewel is pretty old. Even Luminous is getting up there. There have been a lot of multisite improvements in Mimic and now Nautilus, so you might want to consider upgrading all the way to 14.2.4. Anyway, the way we solve this is by giving each zone a different name (eg

[ceph-users] Ceph multi site outage question

2019-10-09 Thread Melzer Pinto
Hello, I have a question about multi site configuration. I have 2 clusters configured in a single realm and zonegroup. One cluster is the master zone and the other the slave. Lets assume the first cluster can be reached at http://application1.something.com and the 2nd one is

Re: [ceph-users] Ceph pg repair clone_missing?

2019-10-09 Thread Marc Roos
Brad, many thanks!!! My cluster has finally HEALTH_OK af 1,5 year or so! :) -Original Message- Subject: Re: Ceph pg repair clone_missing? On Fri, Oct 4, 2019 at 6:09 PM Marc Roos wrote: > > > > >Try something like the following on each OSD that holds a copy of >

Re: [ceph-users] Ceph pg repair clone_missing?

2019-10-08 Thread Brad Hubbard
On Fri, Oct 4, 2019 at 6:09 PM Marc Roos wrote: > > > > >Try something like the following on each OSD that holds a copy of > >rbd_data.1f114174b0dc51.0974 and see what output you get. > >Note that you can drop the bluestore flag if they are not bluestore > >osds and you will need

[ceph-users] Space reclamation after rgw pool removal

2019-10-08 Thread George Shuklin
Hello. I've created an rgw installation, had uploaded about 60M files into a single bucket. Removal had looked as a long adventure, so I "ceph osd pool rm'ed" both default.rgw.data and default.rgw.index. Now I have this: # rados lspools .rgw.root default.rgw.control default.rgw.meta

[ceph-users] Ceph Negative Objects Number

2019-10-08 Thread Lazuardi Nasution
Hi, I get following weird negative objects number on tiering. Why is this happening? How to get back to normal? Best regards, [root@management-a ~]# ceph df detail GLOBAL: SIZE AVAIL RAW USED %RAW USED OBJECTS 446T 184T 261T 58.62 22092k POOLS:

Re: [ceph-users] cephfs 1 large omap objects

2019-10-08 Thread Paul Emmerich
Hi, the default for this warning changed recently (see other similar threads on the mailing list), it was 2 million before 14.2.3. I don't think the new default of 200k is a good choice, so increasing it is a reasonable work-around. Paul -- Paul Emmerich Looking for help with your Ceph

Re: [ceph-users] ceph stats on the logs

2019-10-08 Thread Eugen Block
Hi, there is also /var/log/ceph/ceph.log on the MONs, it has the stats you're asking for. Does this answer your question? Regards, Eugen Zitat von nokia ceph : Hi Team, With default log settings , the ceph stats will be logged like cluster [INF] pgmap v30410386: 8192 pgs: 8192

[ceph-users] ceph stats on the logs

2019-10-08 Thread nokia ceph
Hi Team, With default log settings , the ceph stats will be logged like cluster [INF] pgmap v30410386: 8192 pgs: 8192 active+clean; 445 TB data, 1339 TB used, 852 TB / 2191 TB avail; 188 kB/s rd, 217 MB/s wr, 1618 op/s Jewel : on mon logs Nautilus : on mgr logs Luminous : not able to view

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-08 Thread Philippe D'Anjou
Hi,unfortunately it's single mon, because we had major outage on this cluster and it's just being used to copy off data now. We werent able to add more mons because once a second mon was added it crashed the first one (there's a bug tracker ticket). I still have old rocksdb files before I ran

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-07 Thread Gregory Farnum
On Sun, Oct 6, 2019 at 1:08 AM Philippe D'Anjou wrote: > > I had to use rocksdb repair tool before because the rocksdb files got > corrupted, for another reason (another bug possibly). Maybe that is why now > it crash loops, although it ran fine for a day. Yeah looks like it lost a bit of

Re: [ceph-users] [Nfs-ganesha-devel] 2.7.3 with CEPH_FSAL Crashing

2019-10-07 Thread Daniel Gryniewicz
Client::fill_statx() is a fairly large function, so it's hard to know what's causing the crash. Can you get line numbers from your backtrace? Daniel On 10/7/19 9:59 AM, David C wrote: Hi All Further to my previous messages, I upgraded to libcephfs2-14.2.2-0.el7.x86_64 as suggested and

Re: [ceph-users] Unexpected increase in the memory usage of OSDs

2019-10-07 Thread Vladimir Brik
> Do you have statistics on the size of the OSDMaps or count of them > which were being maintained by the OSDs? No, I don't think so. How can I find this information? Memory consumption started to climb again: https://icecube.wisc.edu/~vbrik/graph-3.png Some more info (not sure if relevant or

Re: [ceph-users] [Nfs-ganesha-devel] 2.7.3 with CEPH_FSAL Crashing

2019-10-07 Thread David C
Hi All Further to my previous messages, I upgraded to libcephfs2-14.2.2-0.el7.x86_64 as suggested and things certainly seem a lot more stable, I have had some crashes though, could someone assist in debugging this latest crash please? (gdb) bt #0 0x7fce4e9fc1bb in Client::fill_statx(Inode*,

Re: [ceph-users] rgw: multisite support

2019-10-07 Thread M Ranga Swami Reddy
Thank you...Let me confirm the same..and update here. On Sat, Oct 5, 2019 at 12:27 AM wrote: > Swami; > > For 12.2.11 (Luminous), the previously linked document would be: > > https://docs.ceph.com/docs/luminous/radosgw/multisite/#migrating-a-single-site-system-to-multi-site > > Thank you, > >

Re: [ceph-users] cephfs 1 large omap objects

2019-10-06 Thread Nigel Williams
I've adjusted the threshold: ceph config set osd osd_deep_scrub_large_omap_object_key_threshold 35 Colleague suggested that this will take effect on the next deep-scrub. Is the default of 200,000 too small? will this be adjusted in future releases or is it meant to be adjusted in some

Re: [ceph-users] cephfs 1 large omap objects

2019-10-06 Thread Nigel Williams
I followed some other suggested steps, and have this: root@cnx-17:/var/log/ceph# zcat ceph-osd.178.log.?.gz|fgrep Large 2019-10-02 13:28:39.412 7f482ab1c700 0 log_channel(cluster) log [WRN] : Large omap object found. Object: 2:654134d2:::mds0_openfiles.0:head Key count: 306331 Size (bytes):

[ceph-users] cephfs 1 large omap objects

2019-10-06 Thread Nigel Williams
Out of the blue this popped up (on an otherwise healthy cluster): HEALTH_WARN 1 large omap objects LARGE_OMAP_OBJECTS 1 large omap objects 1 large objects found in pool 'cephfs_metadata' Search the cluster log for 'Large omap object found' for more details. "Search the cluster log" is

[ceph-users] Hidden Objects

2019-10-06 Thread Lazuardi Nasution
Hi, On inspecting new installed cluster (Nautilus), I find following result. ssd-test pool is cache pool for hdd-test pool. After running some RBD bench and delete all rbd images used for benchmarking, it there is some hidden objects inside both pools (except rbd_directory, rbd_info and

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-06 Thread Philippe D'Anjou
I had to use rocksdb repair tool before because the rocksdb files got corrupted, for another reason (another bug possibly). Maybe that is why now it crash loops, although it ran fine for a day.What is meant with "turn it off and rebuild from remainder"? Am Samstag, 5. Oktober 2019,

Re: [ceph-users] Panic in kernel CephFS client after kernel update

2019-10-05 Thread Kenneth Van Alstyne
Thanks! I’ll remove my patch from my local build of the 4.19 kernel and upgrade to 4.19.77. Appreciate the quick fix. Thanks, -- Kenneth Van Alstyne Systems Architect M: 228.547.8045 15052 Conference Center Dr, Chantilly, VA 20151 perspecta On Oct 5, 2019, at 7:29 AM, Ilya Dryomov

Re: [ceph-users] Panic in kernel CephFS client after kernel update

2019-10-05 Thread Ilya Dryomov
On Tue, Oct 1, 2019 at 9:12 PM Jeff Layton wrote: > > On Tue, 2019-10-01 at 15:04 -0400, Sasha Levin wrote: > > On Tue, Oct 01, 2019 at 01:54:45PM -0400, Jeff Layton wrote: > > > On Tue, 2019-10-01 at 19:03 +0200, Ilya Dryomov wrote: > > > > On Tue, Oct 1, 2019 at 6:41 PM Kenneth Van Alstyne > >

Re: [ceph-users] MDS Stability with lots of CAPS

2019-10-05 Thread Patrick Donnelly
On Wed, Oct 2, 2019 at 9:48 AM Stefan Kooman wrote: > According to [1] there are new parameters in place to have the MDS > behave more stable. Quoting that blog post "One of the more recent > issues weve discovered is that an MDS with a very large cache (64+GB) > will hang during certain recovery

Re: [ceph-users] Unexpected increase in the memory usage of OSDs

2019-10-04 Thread Gregory Farnum
Do you have statistics on the size of the OSDMaps or count of them which were being maintained by the OSDs? I'm not sure why having noout set would change that if all the nodes were alive, but that's my bet. -Greg On Thu, Oct 3, 2019 at 7:04 AM Vladimir Brik wrote: > > And, just as unexpectedly,

Re: [ceph-users] mon sudden crash loop - pinned map

2019-10-04 Thread Gregory Farnum
Hmm, that assert means the monitor tried to grab an OSDMap it had on disk but it didn't work. (In particular, a "pinned" full map which we kept around after trimming the others to save on disk space.) That *could* be a bug where we didn't have the pinned map and should have (or incorrectly

Re: [ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread Maged Mokhtar
The 4M throughput numbers you see now ( 150 MB/s read, 60 MB/s write) are probably limited by your 1G network, and can probably go higher if you increase it ( 10G or use active bonds). In real life, the applications and wokloads determine the block size, io depths, whether it is sequential

Re: [ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread Petr Bena
Thank you guys, I changed FIO parameters and it looks far better now - reading about 150MB/s, writing over 60MB/s Now, the question is, what could I change in my setup to make it this fast - the RBD is used as LVM PV for a VG shared between Xen hypervisors, this is the PV:   --- Physical

Re: [ceph-users] rgw: multisite support

2019-10-04 Thread DHilsbos
Swami; For 12.2.11 (Luminous), the previously linked document would be: https://docs.ceph.com/docs/luminous/radosgw/multisite/#migrating-a-single-site-system-to-multi-site Thank you, Dominic L. Hilsbos, MBA Director – Information Technology Perform Air International Inc.

Re: [ceph-users] ssd requirements for wal/db

2019-10-04 Thread Stijn De Weirdt
hi all, maybe to clarify a bit, e.g. https://indico.cern.ch/event/755842/contributions/3243386/attachments/1784159/2904041/2019-jcollet-openlab.pdf clearly shows that the db+wal disks are not saturated, but we are wondering what is really needed/acceptable wrt throughput and latency (eg is a

Re: [ceph-users] ssd requirements for wal/db

2019-10-04 Thread Vitaliy Filippov
WAL/DB isn't "read intensive". It's more "write intensive" :) use server SSDs with capacitors to get adequate write performance. Hi all, We are thinking about putting our wal/db of hdds/ on ssds. If we would put the wal of 4 HDDS on 1 SSD as recommended, what type of SSD would suffice? We

Re: [ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread Maged Mokhtar
The tests are measuring differing things, and fio test result of 1.5 MB/s is not bad. The rados write bench uses by default 4M block size and does 16 threads and is random in nature, you can change the block size and thread count. The dd command uses by default 512 block size and and 1

Re: [ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread JC Lopez
Hi, your RBD bench and RADOS bench use by default 4MB IO request size while your FIO is configured for 4KB IO request size. If you want to compare apple 2 apple (bandwidth) you need to change the FIO IO request size to 4194304. Plus, you tested a sequential workload with RADOS bench but

Re: [ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread Petr Bena
Hello, I tried to use FIO on RBD device I just created and writing is really terrible (around 1.5MB/s) [root@ceph3 tmp]# fio test.fio rbd_iodepth32: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=rbd, iodepth=32 fio-3.7 Starting 1 process Jobs: 1 (f=1):

Re: [ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread Alexandre DERUMIER
Hi, >>dd if=/dev/zero of=/dev/rbd0 writes at 5MB/s - you are testing with a single thread/iodepth=1 sequentially here. Then only 1 disk at time, and you have network latency too. rados bench is doing 16 concurrent write. Try to test with fio for example, with bigger iodepth, small block/big

[ceph-users] Optimizing terrible RBD performance

2019-10-04 Thread Petr Bena
Hello, If this is too long for you, TL;DR; section on the bottom I created a CEPH cluster made of 3 SuperMicro servers, each with 2 OSD (WD RED spinning drives) and I would like to optimize the performance of RBD, which I believe is blocked by some wrong CEPH configuration, because from my

[ceph-users] ssd requirements for wal/db

2019-10-04 Thread Kenneth Waegeman
Hi all, We are thinking about putting our wal/db of hdds/ on ssds. If we would put the wal of 4 HDDS on 1 SSD as recommended, what type of SSD would suffice? We were thinking of using SATA Read Intensive 6Gbps 1DWPD SSDs. Does someone has some experience with this configuration? Would we

Re: [ceph-users] rgw: multisite support

2019-10-04 Thread Joachim Kraftmayer
Maybe this will help you: https://docs.ceph.com/docs/master/radosgw/multisite/#migrating-a-single-site-system-to-multi-site ___ Clyso GmbH Am 03.10.2019 um 13:32 schrieb M Ranga Swami Reddy: Thank you. Do we have a quick document to do this migration? Thanks

Re: [ceph-users] how to set osd_crush_initial_weight 0 without restart any service

2019-10-04 Thread Paul Mezzanini
That would accomplish what you are looking for, yes. Keep in mind that with norebalance that won't stop NEW data from landing there. It will only keep old data from migrating in. This shouldn't pose too much of an issue for most use cases. -- Paul Mezzanini Sr Systems Administrator /

[ceph-users] mon sudden crash loop - pinned map

2019-10-04 Thread Philippe D'Anjou
Hi,our mon is acting up all of a sudden and dying in crash loop with the following: 2019-10-04 14:00:24.339583 lease_expire=0.00 has v0 lc 4549352     -3> 2019-10-04 14:00:24.335 7f6e5d461700  5 mon.km-fsn-1-dc4-m1-797678@0(leader).paxos(paxos active c 4548623..4549352) is_readable = 1 -

Re: [ceph-users] Ceph pg repair clone_missing?

2019-10-04 Thread Marc Roos
> >Try something like the following on each OSD that holds a copy of >rbd_data.1f114174b0dc51.0974 and see what output you get. >Note that you can drop the bluestore flag if they are not bluestore >osds and you will need the osd stopped at the time (set noout). Also >note,

Re: [ceph-users] Ceph pg repair clone_missing?

2019-10-03 Thread Brad Hubbard
On Thu, Oct 3, 2019 at 6:46 PM Marc Roos wrote: > > > > >> > >> I was following the thread where you adviced on this pg repair > >> > >> I ran these rados 'list-inconsistent-obj'/'rados > >> list-inconsistent-snapset' and have output on the snapset. I tried > to > >> extrapolate your

Re: [ceph-users] NFS

2019-10-03 Thread Marc Roos
Thanks Matt! Really useful configs. I am still on luminous, so I can forget about this now :( I will try when I am nautilus, I have already updated my configuration. However it is interesting that in the configuration nowhere the tenant is specified, so I guess that is being extracted from

Re: [ceph-users] NFS

2019-10-03 Thread Daniel Gryniewicz
"Path" is either "/" to indicate the top of the tree, or a bucket name to indicate a limited export for a single bucket. It's not related to the user at all. On Thu, Oct 3, 2019 at 10:34 AM Marc Roos wrote: > > > How should a multi tenant RGW config look like, I am not able get this > working:

Re: [ceph-users] NFS

2019-10-03 Thread Matt Benjamin
Hi Mark, Here's an example that should work--userx and usery are RGW users created in different tenants, like so: radosgw-admin --tenant tnt1 --uid userx --display-name "tnt1-userx" \ --access_key "userxacc" --secret "test123" user create radosgw-admin --tenant tnt2 --uid usery

Re: [ceph-users] NFS

2019-10-03 Thread Nathan Fish
We have tried running nfs-ganesha (2.7 - 2.8.1) with FSAL_CEPH backed by a Nautilus CephFS. Performance when doing metadata operations (ie anything with small files) is very slow. On Thu, Oct 3, 2019 at 10:34 AM Marc Roos wrote: > > > How should a multi tenant RGW config look like, I am not able

Re: [ceph-users] NFS

2019-10-03 Thread Marc Roos
How should a multi tenant RGW config look like, I am not able get this working: EXPORT { Export_ID=301; Path = "test:test3"; #Path = "/"; Pseudo = "/rgwtester"; Protocols = 4; FSAL { Name = RGW; User_Id = "test$tester1";

Re: [ceph-users] Unexpected increase in the memory usage of OSDs

2019-10-03 Thread Vladimir Brik
And, just as unexpectedly, things have returned to normal overnight https://icecube.wisc.edu/~vbrik/graph-1.png The change seems to have coincided with the beginning of Rados Gateway activity (before, it was essentially zero). I can see nothing in the logs that would explain what happened

Re: [ceph-users] NFS

2019-10-03 Thread Matt Benjamin
RGW NFS can support any NFS style of authentication, but users will have the RGW access of their nfs-ganesha export. You can create exports with disjoint privileges, and since recent L, N, RGW tenants. Matt On Tue, Oct 1, 2019 at 8:31 AM Marc Roos wrote: > > I think you can run into problems

Re: [ceph-users] NFS

2019-10-03 Thread Daniel Gryniewicz
So, Ganesha is an NFS gateway, living in userspace. It provides access via NFS (for any NFS client) to a number of clustered storage systems, or to local filesystems on it's host. It can run on any system that has access to the cluster (ceph in this case). One Ganesha instance can serve quite a

Re: [ceph-users] rgw: multisite support

2019-10-03 Thread M Ranga Swami Reddy
Thank you. Do we have a quick document to do this migration? Thanks Swami On Thu, Oct 3, 2019 at 4:38 PM Paul Emmerich wrote: > On Thu, Oct 3, 2019 at 12:03 PM M Ranga Swami Reddy > wrote: > > > > Below url says: "Switching from a standalone deployment to a multi-site > replicated deployment

Re: [ceph-users] rgw: multisite support

2019-10-03 Thread Paul Emmerich
On Thu, Oct 3, 2019 at 12:03 PM M Ranga Swami Reddy wrote: > > Below url says: "Switching from a standalone deployment to a multi-site > replicated deployment is not supported. > https://docs.openstack.org/project-deploy-guide/charm-deployment-guide/latest/app-rgw-multisite.html this is wrong,

Re: [ceph-users] ceph pg repair fails...?

2019-10-03 Thread Jake Grimmett
Dear All, Many thanks to Brad and Mattia for good advice. I was away for two days, in the meantime the pg has fixed itself. I'm not complaining, but it's strange... Looking at the OSD logs, we see the previous repair fail. Then a routine scrub appears to fix the issue. The same thing happened

Re: [ceph-users] rgw: multisite support

2019-10-03 Thread M Ranga Swami Reddy
Below url says: "Switching from a standalone deployment to a multi-site replicated deployment is not supported. https://docs.openstack.org/project-deploy-guide/charm-deployment-guide/latest/app-rgw-multisite.html Please advise. On Thu, Oct 3, 2019 at 3:28 PM M Ranga Swami Reddy wrote: > Hi, >

[ceph-users] rgw: multisite support

2019-10-03 Thread M Ranga Swami Reddy
Hi, Iam using the 2 ceph clusters in diff DCs (away by 500 KM) with ceph 12.2.11 version. Now, I want to setup rgw multisite using the above 2 ceph clusters. is it possible? if yes, please share good document to do the same. Thanks Swami ___ ceph-users

Re: [ceph-users] rgw S3 lifecycle cannot keep up

2019-10-03 Thread Christian Pedersen
Thank you Robin. Looking at the video it doesn't seem like a fix is anywhere near ready. Am I correct in concluding that Ceph is not the right tool for my use-case? Cheers, Christian On Oct 3 2019, at 6:07 am, Robin H. Johnson wrote: > On Wed, Oct 02, 2019 at 01:48:40PM +0200, Christian

<    5   6   7   8   9   10   11   12   13   14   >