Re: [ceph-users] SLES Packages

2015-06-02 Thread Lars Marowsky-Bree
On 2015-06-01T15:41:05, Steffen Weißgerber wrote: > Hi, > > I'm searching for actual packages for SLES11 SP3. > > Via SMT-Updateserver it seems that there's only Version 0.80.8 available. Are > there > other package sources available (at least for Giant)? Hi Steffen, we have only released th

Re: [ceph-users] Ceph RBD and Cephfuse

2015-06-02 Thread Lars Marowsky-Bree
On 2015-06-02T15:40:54, gjprabu wrote: > Hi Team, > > We are newly using ceph with two OSD and two clients, our requirement > is when we write date through clients it should see in another client also, > storage is mounted using rbd because we running git clone with large amount > of s

Re: [ceph-users] Ceph RBD and Cephfuse

2015-06-02 Thread Lars Marowsky-Bree
On 2015-06-02T17:23:58, gjprabu wrote: > Hi Lars, > >We installed centos in client machines with kernel version is 3.10 > which is rbd supported modules. Now installed ocsfs2-tools and formated but > mount through error. Please check below. You need to configure the ocfs2 cluster

Re: [ceph-users] External XFS Filesystem Journal on OSD

2015-06-04 Thread Lars Marowsky-Bree
On 2015-06-04T12:42:42, David Burley wrote: > Are there any safety/consistency or other reasons we wouldn't want to try > using an external XFS log device for our OSDs? I realize if that device > fails the filesystem is pretty much lost, but beyond that? I think with the XFS journal on the same

Re: [ceph-users] External XFS Filesystem Journal on OSD

2015-07-10 Thread Lars Marowsky-Bree
On 2015-07-09T14:05:55, David Burley wrote: > Converted a few of our OSD's (spinners) over to a config where the OSD > journal and XFS journal both live on an NVMe drive (Intel P3700). The XFS > journal might have provided some very minimal performance gains (3%, > maybe). Given the low gains, we

Re: [ceph-users] External XFS Filesystem Journal on OSD

2015-07-10 Thread Lars Marowsky-Bree
On 2015-07-10T15:20:23, Jacek Jarosiewicz wrote: > We have tried both - you can see performance gain, but we finally went > toward ceph cache tier. It's much more flexible and gives similar gains in > terms of performance. > > Downside to bcache is that you can't use it on a drive that already h

Re: [ceph-users] iSCSI over RDB is a good idea ?

2015-11-05 Thread Lars Marowsky-Bree
On 2015-11-04T14:30:56, Hugo Slabbert wrote: > Sure. My post was not intended to say that iSCSI over RBD is *slow*, just > that it scales differently than native RBD client access. > > If I have 10 OSD hosts with a 10G link each facing clients, provided the OSDs > can saturate the 10G links, I

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-04 Thread Lars Marowsky-Bree
On 2016-07-01T13:04:45, mq wrote: Hi MQ, perhaps the upstream list is not the best one to discuss this. SUSE includes adjusted backports for the iSCSI functionality that upstream does not; very few people here are going to be intimately familiar with the code you're running. If you're evaluating

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-04 Thread Lars Marowsky-Bree
On 2016-07-01T17:18:19, Christian Balzer wrote: > First off, it's somewhat funny that you're testing the repackaged SUSE > Ceph, but asking for help here (with Ceph being owned by Red Hat). *cough* Ceph is not owned by RH. RH acquired the InkTank team and the various trademarks, that's true (and

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-04 Thread Lars Marowsky-Bree
On 2016-07-01T19:11:34, Nick Fisk wrote: > To summarise, > > LIO is just not working very well at the moment because of the ABORT Tasks > problem, this will hopefully be fixed at some point. I'm not sure if SUSE > works around this, but see below for other pain points with RBD + ESXi + iSCSI

Re: [ceph-users] risk mitigation in 2 replica clusters

2017-06-29 Thread Lars Marowsky-Bree
On 2017-06-22T00:51:38, Blair Bethwaite wrote: > I'm doing some work to evaluate the risks involved in running 2r storage > pools. On the face of it my naive disk failure calculations give me 4-5 > nines for a 2r pool of 100 OSDs (no copyset awareness, i.e., secondary disk > failure based purely

Re: [ceph-users] v12.1.0 Luminous RC released

2017-06-29 Thread Lars Marowsky-Bree
On 2017-06-26T11:28:36, Ashley Merrick wrote: > With the EC Overwite support, if currently running behind a cache tier in > Jewel will the overwrite still be of benefit through the cache tier and > remove the need to promote the full block to make any edits? > > Or we better totally removing t

Re: [ceph-users] dropping filestore+btrfs testing for luminous

2017-07-05 Thread Lars Marowsky-Bree
On 2017-06-30T16:48:04, Sage Weil wrote: > > Simply disabling the tests while keeping the code in the distribution is > > setting up users who happen to be using Btrfs for failure. > > I don't think we can wait *another* cycle (year) to stop testing this. > > We can, however, > > - prominentl

Re: [ceph-users] upgrade procedure to Luminous

2017-07-14 Thread Lars Marowsky-Bree
On 2017-07-14T14:12:08, Sage Weil wrote: > > Any thoughts on how to mitigate this, or on whether I got this all wrong and > > am missing a crucial detail that blows this wall of text away, please let me > > know. > I don't know; the requirement that mons be upgraded before OSDs doesn't > seem th

Re: [ceph-users] upgrade procedure to Luminous

2017-07-14 Thread Lars Marowsky-Bree
On 2017-07-14T10:34:35, Mike Lowe wrote: > Having run ceph clusters in production for the past six years and upgrading > from every stable release starting with argonaut to the next, I can honestly > say being careful about order of operations has not been a problem. This requirement did not e

Re: [ceph-users] upgrade procedure to Luminous

2017-07-17 Thread Lars Marowsky-Bree
On 2017-07-14T15:18:54, Sage Weil wrote: > Yes, but how many of those clusters can only upgrade by updating the > packages and rebooting? Our documented procedures have always recommended > upgrading the packages, then restarting either mons or osds first and to > my recollection nobody has c

Re: [ceph-users] [Ceph-maintainers] Ceph release cadence

2017-09-07 Thread Lars Marowsky-Bree
On 2017-09-06T15:23:34, Sage Weil wrote: Hi Sage, thanks for kicking off this discussion - after the L experience, it was on my hot list to talk about too. I do agree that we need predictable releases more than feature-rich releases. Distributors like to plan, but that's not a reason. However,

Re: [ceph-users] [Ceph-maintainers] Mimic timeline

2017-10-05 Thread Lars Marowsky-Bree
On 2017-10-04T21:46:27, Sage Weil wrote: > > After further discussion we are targetting 9 months for Mimic 13.2.0: > > > > - Mar 16, 2018 feature freeze > > - May 1, 2018 release > > > > Upgrades for Mimic will be from Luminous only (we've already made that a > > required stop), but we plan

Re: [ceph-users] who is using nfs-ganesha and cephfs?

2017-11-09 Thread Lars Marowsky-Bree
On 2017-11-08T21:41:41, Sage Weil wrote: > Who is running nfs-ganesha's FSAL to export CephFS? What has your > experience been? > > (We are working on building proper testing and support for this into > Mimic, but the ganesha FSAL has been around for years.) We use it currently, and it works

Re: [ceph-users] RBD device as SBD device for pacemaker cluster

2018-02-06 Thread Lars Marowsky-Bree
On 2018-02-06T13:00:59, Kai Wagner wrote: > I had the idea to use a RBD device as the SBD device for a pacemaker > cluster. So I don't have to fiddle with multipathing and all that stuff. > Have someone already tested this somewhere and can tell how the cluster > reacts on this? SBD should work

Re: [ceph-users] new Open Source Ceph based iSCSI SAN project

2016-10-17 Thread Lars Marowsky-Bree
On 2016-10-17T13:37:29, Maged Mokhtar wrote: Hi Maged, glad to see our patches caught your attention. You're aware that they are being upstreamed by David Disseldorp and Mike Christie, right? You don't have to uplift patches from our backported SLES kernel ;-) Also, curious why you based this o

Re: [ceph-users] new Open Source Ceph based iSCSI SAN project

2016-10-17 Thread Lars Marowsky-Bree
On 2016-10-17T15:31:31, Maged Mokhtar wrote: > This is our first beta version, we do not support cache tiering. We > definitely intend to support it. Cache tiering in Ceph works for this use case. I assume you mean in your UI? Though we all are waiting for Luminous to do away with the need for

Re: [ceph-users] Does anyone know why cephfs do not support EC pool?

2016-10-18 Thread Lars Marowsky-Bree
On 2016-10-18T00:06:57, Erick Perez - Quadrian Enterprises wrote: > Is EC in the roadmap for CEPH? Cant seem to find it. My question is because > "all others" (Nutanix, Hypergrid) do EC storage for VMs as the default way > of storage. It seems EC in ceph (as of Sept 2016) is considered by many >

Re: [ceph-users] new Open Source Ceph based iSCSI SAN project

2016-10-18 Thread Lars Marowsky-Bree
On 2016-10-18T11:36:42, Christian Balzer wrote: > > Cache tiering in Ceph works for this use case. I assume you mean in > > your UI? > May well be, but Oliver suggested that cache-tiering is not supported with > Hammer (0.94.x), which it most certainly is. Right, we've got some success stories o

Re: [ceph-users] Feedback wanted: health warning when standby MDS dies?

2016-10-18 Thread Lars Marowsky-Bree
On 2016-10-18T12:46:48, John Spray wrote: > I've suggested a solution here: > http://tracker.ceph.com/issues/17604 > > This is probably going to be a bit of a subjective thing in terms of > whether people find it useful or find it to be annoying noise, so I'd > be interested in feedback from peo

Re: [ceph-users] New Ceph community manager: Mike Perez

2018-08-29 Thread Lars Marowsky-Bree
On 2018-08-29T01:13:24, Sage Weil wrote: Most excellent! Welcome, Mike! I look forward to working with you. Regards, Lars -- Architect SDS, Distinguished Engineer SUSE Linux GmbH, GF: Felix Imendörffer, Jane Smithard, Graham Norton, HRB 21284 (AG Nürnberg) "Architects should open possib

Re: [ceph-users] Updating standby mds from 12.2.2 to 12.2.4 caused up:active 12.2.2 mds's to suicide

2018-03-14 Thread Lars Marowsky-Bree
On 2018-02-28T02:38:34, Patrick Donnelly wrote: > I think it will be necessary to reduce the actives to 1 (max_mds -> 1; > deactivate other ranks), shutdown standbys, upgrade the single active, > then upgrade/start the standbys. > > Unfortunately this didn't get flagged in upgrade testing. Thank

Re: [ceph-users] Ceph iSCSI is a prank?

2018-03-14 Thread Lars Marowsky-Bree
On 2018-03-02T15:24:29, Joshua Chen wrote: > Dear all, > I wonder how we could support VM systems with ceph storage (block > device)? my colleagues are waiting for my answer for vmware (vSphere 5) and > I myself use oVirt (RHEV). the default protocol is iSCSI. Lean on VMWare to stop being diff

Re: [ceph-users] Updating standby mds from 12.2.2 to 12.2.4 caused up:active 12.2.2 mds's to suicide

2018-03-14 Thread Lars Marowsky-Bree
On 2018-03-14T06:57:08, Patrick Donnelly wrote: > Yes. But the real outcome is not "no MDS [is] active" but "some or all > metadata I/O will pause" -- and there is no avoiding that. During an > MDS upgrade, a standby must take over the MDS being shutdown (and > upgraded). During takeover, metada

[ceph-users] Aligning RBD stripe size with EC chunk size?

2018-06-14 Thread Lars Marowsky-Bree
Hi all, so, I'm wondering right now (with some urgency, ahem) how to make RBD on EC pools faster without resorting to cache tiering. In a replicated pool, we had some success with RBD striping. I wonder if it would be possible to align RBD stripe-unit with the EC chunk size ...? Is that worth p

Re: [ceph-users] Cephalocon Cancelled

2017-05-16 Thread Lars Marowsky-Bree
On 2017-05-15T15:21:47, Danny Al-Gaaf wrote: > What about moving the event to the next OpenStack Summit in Sydney, let > say directly following the Summit. Not sure how many people already plan > to go to Sydney from the Ceph Community, but this way the 10-20 h flight > (depending on your locatio

Re: [ceph-users] Luminous: ETA on LTS production release?

2017-06-19 Thread Lars Marowsky-Bree
On 2017-06-16T20:09:04, Gregory Farnum wrote: > There's a lot going into this release, and things keep popping up. I > suspect it'll be another month or two, but I doubt anybody is capable of > giving a more precise date. :/ The downside of giving up on train > releases... What's still outstandi

Re: [ceph-users] Changing the release cadence

2019-06-26 Thread Lars Marowsky-Bree
On 2019-06-26T14:45:31, Sage Weil wrote: Hi Sage, I think that makes sense. I'd have preferred the Oct/Nov target, but that'd have made Octopus quite short. Unsure whether freezing in December with a release in March is too long though. But given how much people scramble, setting that as a goal

[ceph-users] Erasure Coding performance for IO < stripe_width

2019-07-08 Thread Lars Marowsky-Bree
Morning all, since Luminous/Mimic, Ceph supports allow_ec_overwrites. However, this has a performance impact that looks even worse than what I'd expect from a Read-Modify-Write cycle. https://ceph.com/community/new-luminous-erasure-coding-rbd-cephfs/ also mentions that the small writes would read

Re: [ceph-users] Erasure Coding performance for IO < stripe_width

2019-07-08 Thread Lars Marowsky-Bree
On 2019-07-08T12:25:30, Dan van der Ster wrote: > Is there a specific bench result you're concerned about? We're seeing ~5800 IOPS, ~23 MiB/s on 4 KiB IO (stripe_width 8192) on a pool that could do 3 GiB/s with 4M blocksize. So, yeah, well, that is rather harsh, even for EC. > I would think tha

Re: [ceph-users] Erasure Coding performance for IO < stripe_width

2019-07-08 Thread Lars Marowsky-Bree
On 2019-07-08T14:36:31, Maged Mokhtar wrote: Hi Maged, > Maybe not related, but we find with rbd, random 4k write iops start very low > at first for a new image and then increase over time as we write. If we > thick provision the image it work does not show this. This happens on random > small b

Re: [ceph-users] Erasure Coding performance for IO < stripe_width

2019-07-08 Thread Lars Marowsky-Bree
On 2019-07-08T19:37:13, Paul Emmerich wrote: > object_map can be a bottleneck for the first write in fresh images We're working with CephFS here. -- SUSE Linux GmbH, GF: Felix Imendörffer, Mary Higgins, Sri Rasiah, HRB 21284 (AG Nürnberg) "Architects should open possibilities and not determi

Re: [ceph-users] Ceph features and linux kernel version for upmap

2019-07-10 Thread Lars Marowsky-Bree
On 2019-07-09T16:32:22, Mattia Belluco wrote: > That of course in my case fails with: > Error EPERM: cannot set require_min_compat_client to luminous: 29 > connected client(s) look like jewel (missing 0x800); 1 > connected client(s) look like jewel (missing 0x800); add > -

Re: [ceph-users] What's the best practice for Erasure Coding

2019-07-11 Thread Lars Marowsky-Bree
On 2019-07-09T07:27:28, Frank Schilder wrote: > Small addition: > > This result holds for rbd bench. It seems to imply good performance for > large-file IO on cephfs, since cephfs will split large files into many > objects of size object_size. Small-file IO is a different story. > > The formu

Re: [ceph-users] writable snapshots in cephfs? GDPR/DSGVO

2019-07-11 Thread Lars Marowsky-Bree
On 2019-07-10T09:59:08, Lars Täuber wrote: > Hi everbody! > > Is it possible to make snapshots in cephfs writable? > We need to remove files because of this General Data Protection Regulation > also from snapshots. Removing data from existing WORM storage is tricky, snapshots being a specific

Re: [ceph-users] What's the best practice for Erasure Coding

2019-07-11 Thread Lars Marowsky-Bree
On 2019-07-11T09:46:47, Frank Schilder wrote: > Striping with stripe units other than 1 is something I also tested. I found > that with EC pools non-trivial striping should be avoided. Firstly, EC is > already a striped format and, secondly, striping on top of that with > stripe_unit>1 will ma

Re: [ceph-users] Returning to the performance in a small cluster topic

2019-07-16 Thread Lars Marowsky-Bree
On 2019-07-15T19:08:26, " Drobyshevskiy, Vladimir " wrote: Hi Vladimir, is this a replicated or EC pool? > The cluster itself: > nautilus > 6 nodes, 7 SSD with 2 OSDs per SSD (14 OSDs in overall). You mean 14 OSDs per node, right? > Each node: 2x Intel Xeon E5-2665 v1 (governor = perf

Re: [ceph-users] Returning to the performance in a small cluster topic

2019-07-16 Thread Lars Marowsky-Bree
On 2019-07-16T19:27:32, " Drobyshevskiy, Vladimir " wrote: > This is a replicated pool with size = 3. > Is it possible to check this anyhow? ping shows average latency ~0.147 ms > which is pretty high for IB but might be reasonable (IPoIB). This means that, barring any other overhead, you're li

Re: [ceph-users] New best practices for osds???

2019-07-17 Thread Lars Marowsky-Bree
On 2019-07-17T08:27:46, John Petrini wrote: The main problem we've observed is that not all HBAs can just efficiently and easily pass through disks 1:1. Some of those from a more traditional server background insist on having some form of mapping via RAID. In that case it depends on whether 1 di

Re: [ceph-users] Allocation recommendations for separate blocks.db and WAL

2019-07-18 Thread Lars Marowsky-Bree
On 2019-07-17T11:56:25, Robert LeBlanc wrote: > So, I see the recommendation for 4% of OSD space for blocks.db/WAL and the > corresponding discussion regrading the 3/30/300GB vs 6/60/600GB allocation. > > How does this change when WAL is seperate from blocks.db? > > Reading [0] it seems that 6/

Re: [ceph-users] which tool to use for benchmarking rgw s3, yscb or cosbench

2019-07-22 Thread Lars Marowsky-Bree
On 2019-07-21T23:51:41, Wei Zhao wrote: > Hi: > I found cosbench is a very convenient tool for benchmaring rgw. But > when I read papers , I found YCSB tool, > https://github.com/brianfrankcooper/YCSB/tree/master/s3 . It seems > that this is used for test cloud service , and seems a right too

Re: [ceph-users] Adventures with large RGW buckets [EXT]

2019-08-02 Thread Lars Marowsky-Bree
On 2019-08-01T15:20:19, Matthew Vernon wrote: > One you don't mention is that multipart uploads break during resharding - so > if our users are filling up a bucket with many writers uploading multipart > objects, some of these will fail (rather than blocking) when the bucket is > resharded. Is t

Re: [ceph-users] CephFS snapshot for backup & disaster recovery

2019-08-05 Thread Lars Marowsky-Bree
On 2019-08-04T13:27:00, Eitan Mosenkis wrote: > I'm running a single-host Ceph cluster for CephFS and I'd like to keep > backups in Amazon S3 for disaster recovery. Is there a simple way to > extract a CephFS snapshot as a single file and/or to create a file that > represents the incremental diff

Re: [ceph-users] even number of monitors

2019-08-05 Thread Lars Marowsky-Bree
On 2019-08-05T07:27:39, Alfredo Daniel Rezinovsky wrote: There's no massive problem with even MON counts. As you note, n+2 doesn't really provide added fault tolerance compared to n+1, so there's no win either. That's fairly obvious. Somewhat less obvious - since the failure of any additional M

Re: [ceph-users] AutoScale PG Questions - EC Pool

2019-09-11 Thread Lars Marowsky-Bree
On 2019-09-10T13:36:53, Konstantin Shalygin wrote: > > So I am correct in 2048 being a very high number and should go for > > either 256 or 512 like you said for a cluster of my size with the EC > > Pool of 8+2? > Indeed. I suggest stay at 256. Might as well go to 512, but the 2^n is really impo