Re: [ceph-users] To backport or not to backport

2019-07-04 Thread Daniel Baumann
Hi, On 7/4/19 3:00 PM, Stefan Kooman wrote: > - Only backport fixes that do not introduce new functionality, but addresses > (impaired) functionality already present in the release. ack, and also my full agrement/support for everything else you wrote, thanks. reading in the changelogs about

Re: [ceph-users] Debian Buster builds

2019-06-18 Thread Daniel Baumann
On 6/18/19 3:39 PM, Paul Emmerich wrote: > we maintain (unofficial) Nautilus builds for Buster here: > https://mirror.croit.io/debian-nautilus/ the repository doesn't contain the source packages. just out of curiosity to see what you might have changes, apart from just (re)building the packages..

Re: [ceph-users] Debian Buster builds

2019-06-18 Thread Daniel Baumann
On 6/18/19 3:11 PM, Tobias Gall wrote: > I would like to switch to debian buster and test the upgrade from > luminous but there are currently no ceph releases/builds for buster. shameless plug: we're re-building ceph packages in our repository that we do for our university (and a few other

Re: [ceph-users] Changing the release cadence

2019-06-17 Thread Daniel Baumann
Hi, I didn't bother to create a twitter account just to be able to participate in the poll.. so.. please count me in for October. Regards, Daniel ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Changing the release cadence

2019-06-06 Thread Daniel Baumann
On 6/6/19 9:26 AM, Xiaoxi Chen wrote: > I will vote for November for several reasons: [...] as an academic institution we're aligned by August to July (school year) instead of the January to December (calendar year), so all your reasons (thanks!) are valid for us.. just shifted by 6 months,

Re: [ceph-users] Changing the release cadence

2019-06-05 Thread Daniel Baumann
On 6/5/19 5:57 PM, Sage Weil wrote: > So far the balance of opinion seems to favor a shift to a 12 month > cycle [...] it seems pretty likely we'll make that shift. thanks, much appreciated (from an cluster operating point of view). > Thoughts? GNOME and a few others are doing April and

Re: [ceph-users] Mimic 13.2.3?

2019-01-04 Thread Daniel Baumann
On 01/04/2019 07:32 PM, Peter Woodman wrote: > not to mention that the current released version of mimic (.2) has a > bug that is potentially catastrophic to cephfs, known about for > months, yet it's not in the release notes. would have upgraded and > destroyed data had i not caught a thread on

Re: [ceph-users] Mimic 13.2.3?

2019-01-04 Thread Daniel Baumann
On 01/04/2019 05:07 PM, Matthew Vernon wrote: > how is it still the case that packages are being pushed onto the official > ceph.com repos that people > shouldn't install? We're still on 12.2.5 because of this. Basically every 12.2.x after that had notes on the mailinglist like "don't use, wait

Re: [ceph-users] Problem with CephFS

2018-11-21 Thread Daniel Baumann
Hi, On 11/21/2018 07:04 PM, Rodrigo Embeita wrote: >             Reduced data availability: 7 pgs inactive, 7 pgs down this is your first problem: unless you have all data available again, cephfs will not be back. after that, I would take care about the redundancy next, and get the one missing

Re: [ceph-users] Mimic and Debian 9

2018-10-17 Thread Daniel Baumann
Hi, On 10/17/2018 04:04 PM, John Spray wrote: > If there isn't anything > too hacky involved in the build perhaps your packages could simply be > the official ones? being a Debian Developer, I can upload my backports that I maintain/use at work to e.g. people.debian.org/~daniel or so. Given time

Re: [ceph-users] ls operation is too slow in cephfs

2018-07-17 Thread Daniel Baumann
On 07/17/2018 11:43 AM, Marc Roos wrote: > I had similar thing with doing the ls. Increasing the cache limit helped > with our test cluster same here; additionally we also had to use more than one MDS to get good performance (currently 3 MDS plus 2 stand-by per FS). Regards, Daniel

Re: [ceph-users] fuse vs kernel client

2018-07-09 Thread Daniel Baumann
On 07/09/2018 10:18 AM, Manuel Sopena Ballesteros wrote: > FUSE is supposed to run slower. in our tests with ceph 11.2.x and 12.2.x clusters, cephfs-fuse is always around 10 times slower than kernel cephfs. Regards, Daniel ___ ceph-users mailing list

Re: [ceph-users] samba gateway experiences with cephfs ?

2018-05-24 Thread Daniel Baumann
Hi, On 05/24/2018 02:53 PM, David Disseldorp wrote: >> [ceph_test] >> path = /ceph-kernel >> guest ok = no >> delete readonly = yes >> oplocks = yes >> posix locking = no jftr, we use the following to disable all locking (on samba 4.8.2): oplocks = False level2 oplocks = False kernel

Re: [ceph-users] Ceph replication factor of 2

2018-05-24 Thread Daniel Baumann
Hi, I coudn't agree more, but just to re-emphasize what others already said: the point of replica 3 is not to have extra safety for (human|software|server) failures, but to have enough data around to allow rebalancing the cluster when disks fail. after a certain amount of disks in a

Re: [ceph-users] samba gateway experiences with cephfs ?

2018-05-21 Thread Daniel Baumann
Hi On 05/21/2018 05:38 PM, Jake Grimmett wrote: > Unfortunately we have a large number (~200) of Windows and Macs clients > which need CIFS/SMB access to cephfs. we too, which is why we're (partially) exporting cephfs over samba too, 1.5y in production now. for us, cephfs-over-samba is

Re: [ceph-users] (yet another) multi active mds advise needed

2018-05-18 Thread Daniel Baumann
On 05/19/2018 01:13 AM, Webert de Souza Lima wrote: > New question: will it make any difference in the balancing if instead of > having the MAIL directory in the root of cephfs and the domains's > subtrees inside it, I discard the parent dir and put all the subtress right > in cephfs root? the

Re: [ceph-users] (yet another) multi active mds advise needed

2018-05-18 Thread Daniel Baumann
On 05/18/2018 11:19 PM, Patrick Donnelly wrote: > So, you would want to have a standby-replay > daemon for each rank or just have normal standbys. It will likely > depend on the size of your MDS (cache size) and available hardware. jftr, having 3 active mds and 3 standby-replay resulted May 20217

Re: [ceph-users] Multi-MDS Failover

2018-04-30 Thread Daniel Baumann
On 04/27/2018 07:11 PM, Patrick Donnelly wrote: > The answer is that there may be partial availability from > the up:active ranks which may hand out capabilities for the subtrees > they manage or no availability if that's not possible because it > cannot obtain the necessary locks. additionally:

Re: [ceph-users] Cluster degraded after Ceph Upgrade 12.2.1 => 12.2.2

2018-04-26 Thread Daniel Baumann
ceph is cluster - so reboots aren't an issue (we do set noout during a planed serial reboot of all machines of the cluster). personally i don't think the hassle of live patching is worth it. it's a very gross hack that only works well in very specific niche cases. ceph (as every proper cluster)

Re: [ceph-users] Ubuntu 17.10 or Debian 9.3 + Luminous = random OS hang ?

2018-01-19 Thread Daniel Baumann
Hi, On 01/19/18 14:46, Youzhong Yang wrote: > Just wondering if anyone has seen the same issue, or it's just me. we're using debian with our own backported kernels and ceph, works rock solid. what you're describing sounds more like hardware issues to me. if you don't fully "trust"/have

Re: [ceph-users] CephFS log jam prevention

2017-12-05 Thread Daniel Baumann
Hi, On 12/05/17 17:58, Dan Jakubiec wrote: > Is this is configuration problem or a bug? we had massive problems with both kraken (feb-sept 2017) and luminous (12.2.0), seeing the same behaviour as you. ceph.conf was containing defaults only, except that we had to crank up mds_cache_size and

Re: [ceph-users] ceph-disk is now deprecated

2017-11-30 Thread Daniel Baumann
On 11/30/17 14:04, Fabian Grünbichler wrote: > point is - you should not purposefully attempt to annoy users and/or > downstreams by changing behaviour in the middle of an LTS release cycle, exactly. upgrading the patch level (x.y.z to x.y.z+1) should imho never introduce a behaviour-change,

Re: [ceph-users] CephFS - Mounting a second Ceph file system

2017-11-29 Thread Daniel Baumann
On 11/29/17 00:06, Nigel Williams wrote: > Are their opinions on how stable multiple filesystems per single Ceph > cluster is in practice? we're using a single cephfs in production since february, and switched to three cephfs in september - without any problem so far (running 12.2.1). workload

Re: [ceph-users] CephFS - Mounting a second Ceph file system

2017-11-28 Thread Daniel Baumann
On 11/28/17 15:09, Geoffrey Rhodes wrote: > I'd like to run more than one Ceph file system in the same cluster. > Can anybody point me in the right direction to explain how to mount the > second file system? if you use the kernel client, you can use the mds_namespace option, i.e.: mount -t

Re: [ceph-users] how to debug (in order to repair) damaged MDS (rank)?

2017-10-10 Thread Daniel Baumann
On 10/10/2017 02:10 PM, John Spray wrote: > Yes. worked, rank 6 is back and cephfs up again. thank you very much. > Do a final ls to make sure you got all of them -- it is > dangerous to leave any fragments behind. will do. > BTW opened http://tracker.ceph.com/issues/21749 for the underlying

Re: [ceph-users] how to debug (in order to repair) damaged MDS (rank)?

2017-10-10 Thread Daniel Baumann
Hi John, thank you very much for your help. On 10/10/2017 12:57 PM, John Spray wrote: > A) Do a "rados -p ls | grep "^506\." or similar, to > get a list of the objects done, gives me these: 506. 506.0017 506.001b 506.0019 506.001a 506.001c

[ceph-users] how to debug (in order to repair) damaged MDS (rank)?

2017-10-10 Thread Daniel Baumann
Hi all, unfortunatly I'm still struggling bringing cephfs back up after one of the MDS has been marked "damaged" (see messages from monday). 1. When I mark the rank as "repaired", this is what I get in the monitor log (leaving unrelated leveldb compacting chatter aside): 2017-10-10

Re: [ceph-users] cephfs: how to repair damaged mds rank?

2017-10-09 Thread Daniel Baumann
Hi John, On 10/09/2017 10:47 AM, John Spray wrote: > When a rank is "damaged", that means the MDS rank is blocked from > starting because Ceph thinks the on-disk metadata is damaged -- no > amount of restarting things will help. thanks. > The place to start with the investigation is to find the

Re: [ceph-users] cephfs: how to repair damaged mds rank?

2017-10-09 Thread Daniel Baumann
On 10/09/2017 09:17 AM, Daniel Baumann wrote: > The relevant portion from the ceph-mds log (when starting mds9 which > should then take up rank 6; I'm happy to provide any logs): i've turned up the logging (see attachment).. could it be that we hit this bug here? http://tracker.ceph.com/

[ceph-users] cephfs: how to repair damaged mds rank?

2017-10-09 Thread Daniel Baumann
Hi all, we have a Ceph Cluster (12.2.1) with 9 MDS ranks in multi-mds mode. "out of the blue", rank 6 is marked as damaged (and all other MDS are in state up:resolve) and I can't bring the FS up again. 'ceph -s' says: [...] 1 filesystem is degraded 1 mds daemon damaged