Re: [ceph-users] HEALTH_WARN 1 MDSs report oversized cache

2019-12-05 Thread Ranjan Ghosh
Ah, I understand now. Makes a lot of sense. Well, we have a LOT of small files so that might be the reason. I'll keep an eye on it whether the message shows up again. Thank you! Ranjan Am 05.12.19 um 19:40 schrieb Patrick Donnelly: > On Thu, Dec 5, 2019 at 9:45 AM Ranjan Ghosh wrote: &g

Re: [ceph-users] HEALTH_WARN 1 MDSs report oversized cache

2019-12-05 Thread Ranjan Ghosh
e from the default of 1GiB to 32GiB. My rough > estimate is 2.5kiB per inode in recent use. > > > On Thu, Dec 5, 2019 at 10:39 AM Ranjan Ghosh wrote: >> Okay, now, after I settled the issue with the oneshot service thanks to >> the amazing help of Paul and Richard (thanks again!

Re: [ceph-users] HEALTH_WARN 1 MDSs report oversized cache

2019-12-05 Thread Ranjan Ghosh
Okay, now, after I settled the issue with the oneshot service thanks to the amazing help of Paul and Richard (thanks again!), I still wonder: What could I do about that MDS warning: === health: HEALTH_WARN 1 MDSs report oversized cache === If anybody has any ideas? I tried googling it, of

Re: [ceph-users] What does the ceph-volume@simple-crazyhexstuff SystemD service do? And what to do about oversized MDS cache?

2019-12-05 Thread Ranjan Ghosh
Hi Richard, Ah, I think I understand, now, brilliant. It's *supposed* to do exactly that. Mount it once on boot and then just exit. So everything is working as intended. Great. Thanks Ranjan Am 05.12.19 um 15:18 schrieb Richard: > On 2019-12-05 7:19 AM, Ranjan Ghosh wrote: >>

Re: [ceph-users] What does the ceph-volume@simple-crazyhexstuff SystemD service do? And what to do about oversized MDS cache?

2019-12-05 Thread Ranjan Ghosh
ypted ceph-disk OSDs) > > Paul > > -- > Paul Emmerich > > Looking for help with your Ceph cluster? Contact us at https://croit.io > > croit GmbH > Freseniusstr. 31h > 81247 München > www.croit.io <http://www.croit.io> > Tel: +49 89 1896585 90 > >

[ceph-users] What does the ceph-volume@simple-crazyhexstuff SystemD service do? And what to do about oversized MDS cache?

2019-12-05 Thread Ranjan Ghosh
Hi all, After upgrading to Ubuntu 19.10 and consequently from Mimic to Nautilus, I had a mini-shock when my OSDs didn't come up. Okay, I should have read the docs more closely, I had to do: # ceph-volume simple scan /dev/sdb1 # ceph-volume simple activate --all Hooray. The OSDs came back to

Re: [ceph-users] HEALTH_WARN - 3 modules have failed dependencies

2019-05-01 Thread Ranjan Ghosh
ork at all with the newest Ubuntu version. Only one module can be loaded. Sad :-( Hope this will be fixed soon... Am 30.04.19 um 21:18 schrieb Ranjan Ghosh: Hi my beloved Ceph list, After an upgrade from Ubuntu Cosmic to Ubuntu Disco (and according Ceph packages updated from 13.2.2 to 13.2.4),

[ceph-users] HEALTH_WARN - 3 modules have failed dependencies

2019-04-30 Thread Ranjan Ghosh
Hi my beloved Ceph list, After an upgrade from Ubuntu Cosmic to Ubuntu Disco (and according Ceph packages updated from 13.2.2 to 13.2.4), I now get this when I enter "ceph health": HEALTH_WARN 3 modules have failed dependencies "ceph mgr module ls" only reports those 3 modules enabled:

Re: [ceph-users] Urgent: Reduced data availability / All pgs inactive

2019-02-21 Thread Ranjan Ghosh
pg-states/ /The ceph-mgr hasn’t yet received any information about the PG’s state from an OSD since mgr started up./ ср, 20 февр. 2019 г. в 23:10, Ranjan Ghosh mailto:gh...@pw6.de>>: Hi all, hope someone can help me. After restarting a node of my 2-node-cluster

[ceph-users] Urgent: Reduced data availability / All pgs inactive

2019-02-20 Thread Ranjan Ghosh
Hi all, hope someone can help me. After restarting a node of my 2-node-cluster suddenly I get this: root@yak2 /var/www/projects # ceph -s   cluster:     id: 749b2473-9300-4535-97a6-ee6d55008a1b     health: HEALTH_WARN     Reduced data availability: 200 pgs inactive   services:    

[ceph-users] ceph pg dump

2018-06-14 Thread Ranjan Ghosh
Hi all, we have two small clusters (3 nodes each) called alpha and beta. One node (alpha0/beta0) is on a remote site and only has monitor & manager. The two other nodes (alpha/beta-1/2) have all 4 services and contain the OSDs and are connected via an internal network. In short: alpha0

Re: [ceph-users] Cluster degraded after Ceph Upgrade 12.2.1 => 12.2.2

2018-04-26 Thread Ranjan Ghosh
. if you wish to maintain service during upgrade On 25.04.2018 11:52, Ranjan Ghosh wrote: Thanks a lot for your detailed answer. The problem for us, however, was that we use the Ceph packages that come with the Ubuntu distribution. If you do a Ubuntu upgrade, all packages are upgraded in one go

Re: [ceph-users] Cluster degraded after Ceph Upgrade 12.2.1 => 12.2.2

2018-04-25 Thread Ranjan Ghosh
ntil the osd is up and healthy again. but you have less chance for dataloss then 2/1 pools. if you added a osd on a third host you can run size:3 min_size:2 . the recommended config when you can have both redundancy and high availabillity. kind regards Ronny Aasen On 11.04.2018

Re: [ceph-users] Cluster degraded after Ceph Upgrade 12.2.1 => 12.2.2

2018-04-11 Thread Ranjan Ghosh
Ranjan Am 11.04.2018 um 17:07 schrieb Ranjan Ghosh: Thank you for your answer. Do you have any specifics on which thread you're talking about? Would be very interested to read about a success story, because I fear that if I update the other node that the whole cluster comes down. Am

Re: [ceph-users] Cluster degraded after Ceph Upgrade 12.2.1 => 12.2.2

2018-04-11 Thread Ranjan Ghosh
to update all osd's, mon's etc. I can remember running into similar issue. You should be able to find more about this in mailing list archive. -Original Message- From: Ranjan Ghosh [mailto:gh...@pw6.de] Sent: woensdag 11 april 2018 16:02 To: ceph-users Subject: [ceph-users] Cluster degraded

[ceph-users] Cluster degraded after Ceph Upgrade 12.2.1 => 12.2.2

2018-04-11 Thread Ranjan Ghosh
Hi all, We have a two-cluster-node (with a third "monitoring-only" node). Over the last months, everything ran *perfectly* smooth. Today, I did an Ubuntu "apt-get upgrade" on one of the two servers. Among others, the ceph packages were upgraded from 12.2.1 to 12.2.2. A minor release update,

[ceph-users] Ubuntu upgrade Zesty => Aardvark, Implications for Ceph?

2017-11-13 Thread Ranjan Ghosh
Hi everyone, In January, support for Ubuntu Zesty will run out and we're planning to upgrade our servers to Aardvark. We have a two-node-cluster (and one additional monitoring-only server) and we're using the packages that come with the distro. We have mounted CephFS on the same server with

Re: [ceph-users] Blocked requests problem

2017-08-22 Thread Ranjan Ghosh
flags noscrub,nodeep-scrub,sortbitwise,require_jewel_osds pgmap v17667617: 1408 pgs, 5 pools, 24779 GB data, 6494 kobjects 70497 GB used, 127 TB / 196 TB avail 1407 active+clean 1 active+clean+scrubbing Thanks, Ramazan On 22 Aug

Re: [ceph-users] Blocked requests problem

2017-08-22 Thread Ranjan Ghosh
Hi Ramazan, I'm no Ceph expert, but what I can say from my experience using Ceph is: 1) During "Scrubbing", Ceph can be extremely slow. This is probably where your "blocked requests" are coming from. BTW: Perhaps you can even find out which processes are currently blocking with: ps aux | grep

[ceph-users] WBThrottle

2017-08-22 Thread Ranjan Ghosh
Hi Ceph gurus, I've got the following problem with our Ceph installation (Jewel): There are various websites served from the CephFS mount. Sometimes, when I copy many new (large?) files onto this mount, it seems that after a certain delay, everything grinds to a halt. No websites are served;

Re: [ceph-users] num_caps

2017-05-15 Thread Ranjan Ghosh
the actually performance impact of all those caps is not that bad... :-/ Am 15.05.2017 um 14:49 schrieb John Spray: On Mon, May 15, 2017 at 1:36 PM, Henrik Korkuc <li...@kirneh.eu> wrote: On 17-05-15 13:40, John Spray wrote: On Mon, May 15, 2017 at 10:40 AM, Ranjan Ghosh <gh...@pw6.

[ceph-users] num_caps

2017-05-15 Thread Ranjan Ghosh
Hi all, When I run "ceph daemon mds. session ls" I always get a fairly large number for num_caps (200.000). Is this normal? I thought caps are sth. like open/locked files meaning a client is holding a cap on a file and no other client can access it during this time. How can I debug this if

[ceph-users] Unsolved questions

2017-02-06 Thread Ranjan Ghosh
Hi everyone, I'm now running our two-node mini-cluster for some months. OSD, MDS and Monitor is running on both nodes. Additionally there is a very small third node which is only running a third monitor but no MDS/OSD. On both main servers, CephFS is mounted via FSTab/Kernel driver. The

Re: [ceph-users] Ceph Very Small Cluster

2016-10-24 Thread Ranjan Ghosh
own. I can't recall the exact config options off-hand, but it's something like "mon osd min down reports". Search the docs for that. :) -Greg On Thursday, September 29, 2016, Peter Maloney <peter.malo...@brockmann-consult.de <mailto:peter.malo...@brockmann-consult.de>>

Re: [ceph-users] Ceph Very Small Cluster

2016-09-29 Thread Ranjan Ghosh
Wow. Amazing. Thanks a lot!!! This works. 2 (hopefully) last questions on this issue: 1) When the first node is coming back up, I can just call "ceph osd up 0" and Ceph will start auto-repairing everything everything, right? That is, if there are e.g. new files that were created during the

Re: [ceph-users] Ceph Very Small Cluster

2016-09-29 Thread Ranjan Ghosh
Hi Vasu, thank you for your answer. Yes, all the pools have min_size 1: root@uhu2 /scripts # ceph osd lspools 0 rbd,1 cephfs_data,2 cephfs_metadata, root@uhu2 /scripts # ceph osd pool get cephfs_data min_size min_size: 1 root@uhu2 /scripts # ceph osd pool get cephfs_metadata min_size min_size:

[ceph-users] Ceph Very Small Cluster

2016-09-28 Thread Ranjan Ghosh
Hi everyone, Up until recently, we were using GlusterFS to have two web servers in sync so we could take one down and switch back and forth between them - e.g. for maintenance or failover. Usually, both were running, though. The performance was abysmal, unfortunately. Copying many small files